We are using the solution in the operations space.
Our primary use case is production monitoring of complex business critical systems. Another use case would be performance testing of critical releases.
Download the Dynatrace Buyer's Guide including reviews and more. Updated: June 2022
Dynatrace is an AI-powered software intelligence monitoring platform that accelerates digital transformation and simplifies cloud complexities. Dynatrace is an entirely automated full-stack solution that provides data and answers about the performance of your applications and deep insight into every transaction throughout every application, including the end-user experience. By modernizing and automating enterprise cloud operations, users can deliver an optimal digital experience with higher quality software to customers faster.
Dynatrace offers an all-in-one automated artificial intelligence solution that brings together application performance, cloud and infrastructure, and digital experience monitoring. Dynatrace accelerates performance-driven results through operations, development, and business teams with a shared metrics platform. In addition, users are provided a full-stack monitoring experience with three patented technologies:
What does Dynatrace offer?
Dynatrace redefines how organizations monitor their digital ecosystems. The solution offers:
Reviews from Real Users
Dynatrace is the only solution that provides answers to organizations based on deep insight into each user, transaction, and organization's environment.
Barry P., a managing performance engineer at Medica Health Plans, writes, "With Dynatrace, we have synthetic checks and real-user monitoring of all of our websites, places where members and providers can interact with us over the web. We monitor the response times of those with Dynatrace, and it's all integrated into one place."
A consultant at a tech service company notes, "A feature that's one of the highlights of Dynatrace is the AI. The second most valuable feature is OneAgent. Between infrastructures, applications, operating systems, you can deploy with just a single agent and can practically install and forget about it."
Audi, Best Buy, LinkedIn, CISCO, Intuit, KRONOS, Scottrade, Wells Fargo, ULTA Beauty, Lenovo, Swarovsk, Nike, Whirlpool, American Express
We are using the solution in the operations space.
Our primary use case is production monitoring of complex business critical systems. Another use case would be performance testing of critical releases.
The solution uses a single agent for automated deployment and discovery, which helps our operations. It reduces the cost of ownership of managing Dynatrace as a tool set, ensuring that we're able to maximize the value from Dynatrace and monitoring is available. That's a big plus.
An example of how it helps is we are more proactive than we were previously, though we're not quite where we want to be. Engineers are talking more with the operations people, which is closing the loop. Our teams are becoming more customer centric.
The platform is very good at identifying potential issues, but each problem that surfaces in most cases still needs to be qualified and quantified by somebody who understands the system. Complex application problems, not infrastructure, surfaced by Dynatrace still need to be reviewed by somebody who understands the application logic or system architecture. For somebody who understands the platform though, issues can resolved in minutes as opposed to hours.
We have the ability to detect user action response time slow downs and their consequences, along with the back-end calls to third-parties. We are heavily dependent, for a number of products, on back-end service calls to other suppliers. Using Dynatrace, we are able to measure the performance of those third-parties.
We are also using Dynatrace to right-size the infrastructure, especially on private cloud where we have to provision the resources upfront to save costs. Dynatrace helps us by finding how many resource we are utilizing and identifies how many resources we need to maintain for the level of performance and scalability that's required. This has helped us right-size in about 50 percent of our cases, leading to a reduction in cloud resources by 50 percent.
The solution helps DevOps to focus on continuous delivery and shift quality issues to pre-production. This helps with performance testing because our performance test teams are more aware of how product features are performing, which helps to prioritize our testing. It creates test cases so we're able to do more testing. Because Dynatrace helps us define the cause more quickly, this speeds up the time between test cycles.
The end-to-end trace is valuable for us to be able to assign responsibility to the right resolver group very quickly.
The user experience allows us to be able to gauge customer experience and understand the performance impact of our platform.
It has a very nice interface with an easy way to visualize the data that we need, making it quickly accessible. It is very easy to use.
As a platform consolidating tool, it covers 90 percent of the needs for most applications. In that respect, it presents a very high value for us.
We have used synthetic monitoring functionalities to poll. Mostly, it's around service availability and key functionality of a website from different geographic locations.
The real-user monitoring is mostly used to gauge the difference in performance for multitenant applications, This is so we can discern if there are any local network or client-facing issues when we do a comparison between each customer. It is quite important for us to be able to identify a client-side issue, as opposed to a feature managed problem, because we're essentially providing managed services of business applications.
Dashboarding and having different templates available for more business reporting, or even other metrics, would be useful.
With Dynatrace, we use one tool where we would have used many, but we still have had gaps.
It has very high availability.
When we started, we were measuring uptime in a different way, and then Dynatrace started measuring uptime based on services, as opposed to infrastructure. Initially, because we started using different metrics for availability, it showed us that we weren't available as much as we thought we were. This helped us to have better conversations with customers and improved availability from the customer perspective over time.
We have 115 users, which includes Level 2 and 3 supports, service design, product management, cloud infrastructure management, software developers, software testers, and product architects.
We are only in an early phase at the moment regarding the use of Dynatrace. Currently, we are only using it on two critical platforms. Going forward, we're looking to expand to nine critical platforms.
Our adoption rate across the portfolio is low because we're still in a pilot phase trying to build out our business cases.
The technical support is excellent and very fast. Not only do I get a quick response, but they're also able to close the request off very quickly and satisfactorily with a fix.
Some of the feedback I get from our team, who are familiar with other tools: "Compared with other tools, Dynatrace support is excellent."
The feedback that I get from people is that the initial setup was very straightforward and easy. It was amazing what information we got in such little time after deploying the agent.
In most cases, the deployment is quick. It takes a couple of hours.
For high-risk applications, which are business critical or high complexity, we would deploy Dynatrace. For medium-risk applications, we would consider using Dynatrace. It comes down to cost qualification for medium-risk applications.
The solution has decreased our mean time to identification. It has saved us from 10 minutes to a couple of hours.
Consider volume because that is where you will get the most benefit. Doing a point solution is not cost-effective.
There are additional Professional Services costs which ensure the solution is configured with meaningful names so you're getting the most money for your investment.
It is the easiest platform to manage in comparison to the competition, like Elastic Stack, New Relic, AppDynamic, Nagios, or Prometheus.
Without a doubt, I'd recommend Dynatrace for business critical applications and anything that's driving revenue.
Biggest lesson learnt: To recognize the most value from the information that Dynatrace provides, you need to make it available to everybody in the DevOps group. There is a wealth of data which can be exposed, manipulated, and consumed by other systems, not just what's visible in Dynatrace. This can also be used for inputs into other upstream platforms.
Understand the demands within your environment and plan a pipeline, then discuss with Dynatrace.
We're aware that there are use cases for notifications that can be used for triggering self-healing or autoscaling, but we are not using those yet.
I would rate this solution as a nine (out of 10).
We're a health plan, a health insurer. We're not a big one, we have about a million members. We are growing through adding new business and we're looking to expand into the government programs: Medicare, Medicaid. Right now we provide individual and family, large corporate, self-insured, and a couple other types of health plans.
We are headquartered in Minnesota, outside of Minneapolis. We have a data center in Minnetonka and one in another suburb. We do most of our work on-premise. We don't have much in the cloud for our core backroom applications. We use a package from a company called HealthEdge in Boston, to do our claims processing, membership, enrollment, etc.
Our main use case is application performance monitoring, right at Dynatrace's sweet spot. First, we wanted to know what the performance of our healthcare and our health claims processing system was. Then we wanted to be able to segment it by where the transaction response time is spent. We also wanted to get into the deep dive of the Java profile, because HealthEdge is a Java application that runs on several JVMs. We wanted not only to get into the Java code but to get into the SQL that's created to call into the database, which is where the response-time problems are.
We're using Dynatrace SaaS now. It's the newest version.
We found an error where developers used a Google API that was supposed to find a Medicare workshop by loading a Google Map and help a member find a place where they could go to a Medicare workshop. The API had so many calls an hour and we saw that, usually, about 45 minutes after the hour, that transaction was failing. It turned out that we'd used the 1,000 allocated calls and, when, the hour turned over, it worked again. It integrates all things monitoring, from an application perspective: synthetic, real users, and Java deep-dive.
Dynatrace provides us with a huge benefit for triage because by the time a Dynatrace problem is open, AI has identified all the components and where the response-time issue is or where the failure is. It's really mindless. We don't have to try to pull out a map and figure out how the application looks.
And Dynatrace has a feature called SmartScape. I don't use it a lot because their AI is so good that I've never had to go dig through it myself. But if I were to go through it, it would go from data centers to hosts to processes to services and applications, to show how they're all linked together. So it has a topology view. We use that sometimes when we're doing performance testing, which is something another part of my team does. They need to know which pieces are involved and this helps them know that.
But from a day-to-day event-management and IT operations-center perspective, the Dynatrace AI is what has identified the failing component. The dashboard has all the problems. They open up these problems, which are already events in our ServiceNow environment, and these problems have the call-path and everything else laid out in them. So I've never had to dig into the Smartscape to figure out where my failure is. The Dynatrace AI has done that for me.
What we found early on in our HealthRules environment was that the response-time problems were, 99 percent of the time, in the type of SQL that we throw at the database, because the DBAs would say, "It's not the database, it's the bad SQL." Dynatrace helped us focus immediately on that and get away from: "Is it the network? Is it the server? Is this too busy?" There are all the different things that the vendor wants to throw at you. I went up to Boston to help the vendor a year or two ago. I took them right through the code and the response times and said, "Here's the piece of SQL that makes this particular function slow." Dynatrace was able to do that. We got there in minutes. They said, "Well, your server might be too busy, it might be your network," and I could say, "No, it's none of that. Here's the response time of that transaction and here is the decomposition of it. The thing runs for 13 seconds and spends 12 seconds on this one piece of SQL. I think that's where your problem is." Dynatrace was a huge help there.
The solution has decreased our MTTR by well over 50 percent and maybe by as much as 90 percent. It enabled us to identify some things, first of all. Before, it was endless war rooms, and not really an identification. Dynatrace has driven that almost to zero. When the problem is opened, we know the root cause.
As for mean time to repair, since we know what we need to repair, we can point the developer right at it. It has decreased that by 50 to 60 percent.
It has also dramatically improved our uptime. One of the biggest problems we have with the JVMs, of course, is garbage collection and memory saturation. A memory leak will develop and Dynatrace will show the memory increasing steadily. It will create a problem and they'll work on the problem proactively, and either fix it or schedule graceful downtime. If they have to shut down the environment, they can stage through the three different servers in a type of HA arrangement. So without any disruption to the client, we've been able to fix things that would have turned into major outages of the whole environment. It's a definite help on the preventive side.
In terms of time to market, the guys who work on our web portal interface, who are in-house, were early adopters of the technology on our team and learned what works and what doesn't. Dynatrace has significantly decreased their time to market. They're not really part of the development cycle, but the way they use it and the things they say about it and the reports they've made indicate that it has probably cut nearly 50 percent of the development of their portal code.
It has also helped us with consolidation of tools. We got rid of some New Relic and we got rid of some older tool which was a great, early innovator in this space, but it was acquired by CA or Microsoft. We were still paying licenses for that and were able to consolidate it. We were about to buy a network tool to help us with ACI conversion on our network side, a tool that would mainly tell us who an application is talking to on the and network. We use Dynatrace to do that, so we saved tens of thousands of dollars in not acquiring that tool. We also took the synthetic work that we paid an outsourced company to do for us and we converted all of that. Once we had Dynatrace in the house, we could do it ourselves and that saved $20,000 to $30,000 a year. There's probably more, if I were to look at it, that I could do with Dynatrace. I have to focus on the core system right now, but I think they'll get it in the SNMP monitoring space soon, if they're not already there. And the plugins on the ActiveGates have a lot of capabilities we could use. We already monitor our VMware environment with it now.
We've started to use the Apdex score in all of our communications. It's a standard metric that's used for websites to indicate how they're performing. That idea is baked into Dynatrace and we've built on that throughout our company. The weekly service quality reports that are produced and sent via email to all Dynatrace users are starting to get some notice. They show, from the web portal side, what the Apdex is. Is it acceptable, tolerating, or unacceptable? It shows the percentages of the time of use and where they're coming from. It also shows it geographically and what type of browser most of your users are using. It shows how much of it is mobile versus desktop, which has proved very valuable to our digital experience people. Things like that are a huge benefit, and those are things I didn't even know existed when I bought it.
In addition to just the monitoring of the HealthRules ecosystem — which is typical BusinessWorks, Oracle Databases, and JVMs for transactions — we do a lot of web monitoring. With Dynatrace, we have synthetic checks and real-user monitoring of all of our websites, places where members and providers can interact with us over the web. We monitor the response times of those with Dynatrace, and it's all integrated into one place. We actively synthetically monitor our websites from two or three geographic locations. Our business is in nine States, so we're not international by any means. We sell health insurance to members in Oklahoma, Kansas, North and South Dakota, Wisconsin, Minnesota. We monitor those synthetically.
It also instruments .NET, and BusinessWorks out-of-the-box.
It has an integration with ServiceNow, which is great. Dynatrace creates tickets for things and its AI finds root cause. We have integrated that with our ServiceNow to generate events and incidents, so that all of our event management will be done in the ServiceNow Developer. We're working on that now. In terms of the self-healing aspect, we don't use Dynatrace to do that, although we could. We've gone down the path of trying to use ServiceNow's Orchestration. But we may come back to Dynatrace for that, depending on how that works.
In addition to ServiceNow, there is a CMDB integration, so when a Dynatrace problem is discovered, the Dynatrace ID correlates to a CMDB and that's how we open an incident or event. We don't need to do the correlation. If an event turns into an incident, then the correlation is done automatically with the Dynatrace ServiceNow application, which is in the ServiceNow store. It syncs up the CMDB's entries, the CIs, with the Dynatrace IDs so that all of the different pieces of the response-time puzzle that Dynatrace has, can be assigned to a CI in our CMDB. We are actively working on improving our discovery in CMDB, as it's not the most robust. Dynatrace is a huge help there because the OneAgent discovers all these things for us. So it helps with ServiceNow discovery as well.
The Dynatrace panel generally lets you know how many users it affects, and how many transactions or events in that application it affects. We don't use that a lot. That's beyond our capability right now, but I don't see any reason why it wouldn't be quite useful to assign severity from that.
Around the way licensing works, I would like to put it everywhere in infrastructure-only mode and I want it to be reasonable to do that.
From a technological standpoint, there is the OneAgent versus plugins they have. They called them security gateways when they first came out. They're the way that the OneAgents talk to local active gates, which communicate out to the Dynatrace cloud to store all the performance data. Instead of every agent going out to the cloud, there's just one spot and security likes that. But they've leveraged those security gateways and renamed them ActiveGates, and now there are different web plugins we can run on it. Sometimes the plugins are designed for things where you put in an agent, Like an Oracle instance of Exadata, or an Oracle appliance. We can't put a OneAgent on that. It's not a standard Linux or Windows OS, so the ActiveGate solution is better there. Sometimes the development of those seems to be running very fast and it's not complete. They don't yet function quite as easily as the OneAgents do. But I have hopes that that's going to get better. We have tried the MQ, the Citrix, and the Oracle ActiveGate plugins. They could be sharper. It's the right direction to go. It just seems like it could be smoother.
I have been using Dynatrace for close to three years in my current company, and before that I used the earlier versions of Dynatrace, DC RUM, at a previous job.
I had one problem early on with WebLogic where Dynatrace was not stable and it would actually affect the ability of one of the WebLogic components. It was instrumented because we thought we needed it to be, but it didn't need to be. When we decided not to instrument it the problem went away.
But that's the only stability issue I've ever had with it. That was the only time it's caused an outage or been responsible for high resource consumption. Typically the OneAgent is well under 1 percent CPU utilization and takes very little memory.
It's used constantly by several teams. They use the Dynatrace mobile app on their phones to get notified of problems in the environment before ServiceNow even notifies them. Our platform services team, which is the team responsible for the HealthEdge environment — if we were a bank, it would be all the backroom functions. It is where you pay claims, enroll members, credential providers and maintain all that stuff. That support team has it on their phones. Our portal team also has the mobile app, so it's used constantly. I hear about it when it's not available, or if there's something odd going on with the mobile app.
It could handle a much larger environment. I add ActiveGates mainly for redundancy. I don't think I need as many as I have. I could scale it out very large. I don't see any limitations. I've never had a problem with that other than my checkbook.
We've tried scaling it to cloud-native environments a little bit. We have a few things that are off-premises, like Microsoft Dynamics and Salesforce, which are in the cloud. We have a cloud-based application that does provider credentialing, as well. We don't have anything that we own in the cloud, so we can't instrument AWS or anything like that with it.
Tech support has generally been pretty good. We get good response. They have a thing called Dynatrace ONE and I find the tech support to be best if I engage it through a chat window on Dynatrace. There's a place, right in the tool, where you can get a hold of a Dynatrace ONE person and they'll look at your problem right away. That seems to work better than the old model of calling support or sending an email, because you would go back and forth. "Send me more doc. What about this? Send me that." The Dynatrace ONE agent gathers everything he needs and, once he has all that, if he doesn't know what the problem is, at something like a level-one triage, he'll open the incident for you and it's done. I like that part. The traditional send-them-an-email, open-a-ticket-online takes too long. The Dynatrace ONE agent available through chat is a great concept. I encourage my team to use that rather than opening a problem. And that's included in the standard licensing.
For our deployment, we did the first 40 in less than an hour. That required a part of one guy, and he maintains it all now. We have close to 200 nodes with OneAgent on them and four ActiveGates, synthetic monitoring, and plugins for MQ and Citrix, among other things. That takes three-fourths of a person on my team. I've federated the support for a lot of the stuff on our portal side. Our portal team developers fell in love with it so much that I just let them run with it and install it as needed. I give them more and more administrative rights. If you add their time, it works out to the equivalent of about a person.
We have close to 100 users. Some of them are just management who use the reports. Some of them are the portal team who are administrators, just like my team, and the majority are in IT. We're starting to take it out to our sales organization, as they're interested in the response time and other things.
We see ROI in performance tuning — improving application performance — big-time. We have teams using it constantly to make our digital experience better, performance-wise and availability-wise. Another part of my group is load testing. They use it as they do their load tests. They use LoadRunner to build a load test and use Dynatrace to monitor after every new release of the HealthRules code to tell them what's better and what's not. There is a huge ROI on load testing and performance testing.
There is also incident response, preventative incident response. We even had the CIO come into my boss's office one day and he was able to say that Dynatrace saw a problem and it was fixed and we didn't have an outage. And he looked at him and said, "That's how it's supposed to work, right?" What the CIO had been promised for 10 years, he finally actually saw an instance of it "in the wild" where we preemptively discovered a problem and fixed it. That's a huge win.
Also, reporting and analytics — to know what the response time is, and how many users use it, just the simple things — are huge.
I'm not sure how to estimate how much Dynatrace has saved us overall. But it's had to have saved us on the order of millions.
We license it for two environments, typically all of production and all of one lower environment, usually our staging environment. If there is a downside to Dynatrace, the only thing I can think of would be the cost. If it were cheaper, I'd have it in all my environments. I don't think they're charging more than it's worth, by any means. It's just that good software costs money.
They have the OneAgent which you buy and install. You can run that in infrastructure-only mode and pay less. The cost is a bit funny, it's calculated based on the memory size of the server you put it on. Sixteen gigabytes of memory, for instance, is one host unit and a host unit costs you, say, $1,000. (I don't recall what the actual cost is, I'd have to look at our contract). There's a switch they've added for infrastructure-only mode, which will cut that cost to about one-sixth or one-seventh of the cost of a full host agent. You won't get the deep-dive response time metrics, but you'll get the infrastructure stuff, which sometimes is all you want.
In addition to the host agent fee, which was the first thing I bought, based on the memory size of the server, the other is in metrics that we collect through the ActiveGate plugins. They charge you per metric.
So the three principle things they charge you for are OneAgent, how many metrics you collect through the ActiveGate, and digital experience monitoring units, or DEM units. Those are basically the cost of the synthetic things, per test. Those things are quite reasonable in cost. The biggest cost is the OneAgent.
The cost to get us up, my first allocation, was under $100,000. My first PO was for about $60,000 and it covered almost our whole production HealthRules environment. We started out with 40 host units and we've grown to 200-plus, and we're a small place. Down the street is a health-related business and I think they have 20,000 host units.
We started by looking at industry reviews and selected the top four or five up in the upper-right quadrant: Dynatrace, AppDynamics, New Relic, and we had a brief look at what at that time was a CA product, or it might've been BMC.
We evaluated the four of them on paper and then brought two in for a trial, a proof of concept: Dynatrace and AppDynamics. Ultimately we selected Dynatrace.
There were several advantages to Dynatrace. Dynatrace was new. Its presence in the cloud was nice, but I could also run it on-prem if I wanted to and, at the time I didn't know which way I was going to go — which way I'd be allowed to go by security. AppDynamics was cloud-only at the time.
For installation, Dynatrace was trivial compared to AppDynamics. AppDynamics had an engineer onsite for two or three weeks and they still couldn't meet all of our use cases, which were pretty simple. I did them first. Then I went to Dynatrace and they said, "Well, download it, install it, and call us If you have any questions." And I thought, "Well, geez, don't I get any hand holding or anything?" It turned out that it was because I didn't need it. It was that simple. You download it, install it, and it injects itself. You can control it. It was just engineered for ease of use, by far. So the installation was night-and-day different.
We have a lot of TIBCO BusinessWorks code around that that we wanted to instrument, and with AppDynamics we had to go into every business process and change the startup. We had hundreds of them and that was a real pain. We had to select which ones and do the work, whereas with Dynatrace, it would discover. Dynatrace has a concept called OneAgent, which you install on the server and it discovers things that you can monitor. You just click on them and say, "I want these monitored," or "Don't monitor these." It takes care of all that work and that was a huge difference. I didn't need a huge staff to maintain it. I didn't need a lot of time from the support teams — because they don't have it — to help me with monitoring. We were able to do the monitoring ourselves.
Then, once it was up and running, the use cases were pretty simple. One was to create a business-level dashboard of response time, and I don't think AppDynamics ever got that out for me.
Dynatrace is easy to use from that perspective. It's easy to install and maintain. I have a small team and one person is my Dynatrace SME, but he does other things as well, so it's not even a full-time job.
I've been doing this for close to 30 years. I've worked for software vendors and I've worked for major companies and now I'm at this small healthcare organization. The "holy grail" has always been the ability to decompose response time and Dynatrace has done that and integrated all of my APM needs in one tool. That is the biggest benefit to me. I can do application performance, from web to Java deep-dive, in one place. That's probably why it costs so much.
If you're thinking about Dynatrace, consider how easy it is to install and maintain. It has broad coverage and it's easy to use. I don't know how the rest of the market even competes anymore; it must be on cost.
As an APM tool, I'd probably rate it at nine out of ten. There are a few rough edges, but I think that's mainly because they're trying to do the right thing too fast.
We have several uses for Dynatrace. Most of the time, we use Dynatrace for looking into potential site problems, investigating reported issues, and trying to replicate those problems in a test environment using the information provided by Dynatrace.
We use Dynatrace for performance monitoring. Quarterly, we will specifically see if there's anything that we can optimize on the front-end of our website, so that's what you see and interact with on the web page.
We also use it to get ahead of any potential problems in our stack. E.g., if Dynatrace is indicating a problem, we will look into it and determine if it's affecting users. Depending on its impact, and usually if it's impacting customers, we can use that information to decide on what we need to work on next to benefit the customer experience.
I use the tool as more of an analyst. I will use Dynatrace to show where systems need to be fixed, etc.
This solution is SaaS. We use Google Cloud Platform, where we just use their compute engines as far as our hosts. We also have a few services that are on-prem. Dynatrace works fine with both of them.
The solution helps our DevOps to focus on continuous delivery and shift quality issues to pre-production. We recently got a staging environment implemented with Dynatrace. We are mainly using it for load testing at the moment. Dynatrace has been detecting failures, letting us know immediately what types of failures are occurring so we can catch them before releases. Our developers have been able to identify bottlenecks and other types of problems that they would not have been able to before by just using standard logging and analytics tools.
The solution give us 360-degree visibility of the user experience across channels, which is a great benefit. We're in eCommerce as a retailer. We are selling across multiple channels and platforms. We have a mobile app and a website. We even have other services which we may instrument with Dynatrace in the future. As far as our website and mobile app that we have instrumented with Dynatrace, it has all been very positive.
The solution has decreased our time to market with new innovations/capabilities because we have been able to quickly identify areas that we can improve for new features and gather that data from Dynatrace. Then, we have been able to verify that our new features and releases are working as expected.
The User Sessions Query language has definitely been the most helpful with its key user actions and user session properties. Using those together, that has completely transformed how we're able to identify customers and their problems on our site. It has made a very big impact over the year.
Using synthetic monitors, we monitor our websites. We have two main domains. There are several plain HTTP monitors, then there are actual browser based monitors that emulate browser behavior. We use both of those types. We have several mobile browsers emulated under synthetic monitors that we use. Those ping our website every 15 minutes. On some of these synthetic monitors, we use multiple data centers to get an idea of geographic availability. We also monitor some of our third-party providers using our synthetic monitors. We monitor our customer support live chat server, which is hosted by a third-party, where we are given alerts if that system were to go down. We are also monitoring an email capture API that's a part of our website.
With user session queries, the main thing that we use that for (and the most valuable), is when we get a problem. If we get some type of a report, obscure problem, or Dynatrace reports a problem, we go straight to using the User Sessions Query Language to find sessions with Session Replay, then we replay those sessions to figure out exactly what the customer did and what conditions may have caused the problem to gauge the impact of the problem itself.
We also save user sessions queries into dashboards, then create different dashboards based on different projects to try and gather data. E.g., last year, we redid a part of our website and used Dynatrace sessions queries and Session Replays to verify that our customers were not having any problems or being confused by their experience. We wanted to verify that, which is one way that we've used the User Sessions Query Language along with the dashboards. We've also created some other dashboards that return custom metrics for us, which goes along, in some cases, with user session properties and user action properties. In that way, we're able to get a very granular look at certain statistics where it would be more difficult to get those numbers from our traditional analytics suite.
The solution’s ability to assess the severity of anomalies based on the actual impact to users and business KPIs is a bit off. I have found that even though Dynatrace detects a problem and gives you a count and estimate of impacted users, this number is usually much higher than is actually the case and not fully accurate. E.g., I recently noticed an error. Every time someone would experience this error, Dynatrace would create a new problem and it would say, "Several hundred people were impacted." However, using Dynatrace's own tools (user Session Replay), then going back and actually tracing through these requests, we found much fewer people were actually impacted. In some sessions that Dynatrace said were impacted, when you view the Session Replay videos, you could see that the customer was not impacted in any meaningful way.
The solution’s ability to visualize, understand our infrastructure, and to do triage is helpful. I wish that you could do user session queries with those host level metrics and be able to create custom graphs the same way you could with user session data. They're both part of Dynatrace, but they don't feel like they're integrated together well. E.g., we're having an issue that has to do with just HTTP codes and we would like to marry that up with a user session query turning that into a dashboard. We can't currently do that because the User Sessions Query Language does not have access to the HTTP errors or HTTP status code data that is part of the hosts and infrastructure package. Otherwise, if you're just focusing on the infrastructure part it, I think it does a good job.
I have been using Dynatrace since February 2019.
I have noticed a few times where data collection did get interrupted. It was two or three times within the past year. Obviously, it's our monitoring system and we don't want that to go down at all. However, three times for no more than 30 minutes each time is pretty good.
The scalability has been able to meet all of our needs. We have not encountered any limitations when scaling Dynatrace with the Google Cloud Platform.
In the past 365 days, we have two websites that we monitor with Dynatrace, including mobile apps. We've recorded over 23 million sessions for Rack Room Shoes and 8.1 million sessions for Off Broadway Shoes.
There are three users who are active users of Dynatrace:
I have noticed a few problems with the service before. I reached out to support and the system did appear to resolve itself on its own (after there was a problem). Then, the support staff couldn't see any further issues. The solution’s self-healing functionality works.
The technical support is below average. They've solved some of the problems that we had, but it took several weeks to resolve almost each problem we had when they probably should have been fixed within a day or two.
There was an initial implementation of AppMon (another Dynatrace offering) before the current Dynatrace SaaS offering.
Dynatrace has definitely made an impact. We were never able to get granular data with any of our other solutions. They were all very disconnected and separate, whereas Dynatrace seems to have good integrations with our entire stack. There haven't been any problems getting additional data now that we have Dynatrace,
It is very easy to use and set up. It did take some customization to get it working for our sites, but after that, it's been pretty easy and straightforward.
The initial setup is complicated, but it's much less complicated than similar systems that I have used in the past. For Dynatrace's setup, maybe there were problems with how our web application was initially developed before I joined Rack Room, because there were a lot of features related to error reporting. It would report errors for things that weren't actual problems, etc. You have to configure it to get around those types of problems, but it's usually fine afterwards.
Over the past year, we've been tweaking Dynatrace. It's been a slow phase-in rollout as far as how much we rely on the data it's giving us back.
I was involved in the initial implementation.
The solution has decreased our mean time to identification by about three days.
The solution decreased our mean time to repair by around a week.
There has been a huge increase in uptime. It's hard to say by how much for certain because we've made other development practice changes.
It is a great platform. We found a lot of value in setting up user session properties and user action properties, then being able to use them to identify individual problems/customers. We use that to sort of streamline the whole process of finding and fixing problems.
Biggest lesson learnt: Customers do not always behave as expected.
I would rate Dynatrace as an eight (out of 10).
We use it to follow up user experience data. It's all banking applications. For example, when you're viewing your account, you open up your mobile app and the click you do to view your account is measured in Dynatrace. It's stored and we are checking the timing at each moment.
We are also following up the timing differences between our different releases. When we have a new version release, we are already checking within our test environment to see what the impact of each change is before it goes to production. And we follow that up in production as well.
In addition, we are following up the availability of all our different systems.
And root cause analysis is also one of the main business cases.
So we have three main use cases:
We use the on-prem solution, but it's the same as the SaaS solution that they are offering. They have Dynatrace SaaS and Dynatrace Managed, and our is the Managed. Currently we're on version 181, but that changes every month.
The dynamic microservices for Kubernetes is really value-added because there is a lot of monitoring functionality already built into Kubernetes Docker. There are also free things like Prometheus which can display that. That's very good for technical people. For the owner of the pod itself, that's enough. But those things don't provide any business value. If you want business value from it, you need to extract it to a higher level, and that's where you need the correlations. You need to correlate what is between all these different services. What is the flow like between the services? How are they interconnected? And that's where Dynatrace gives added value. And the fact is that you can combine these data, which are coming from Kubernetes, and include them in Dynatrace, meaning you have a single pane of glass where you can see everything. You can see the technical things, but you have the bigger business value on top of it, as well.
Before Dynatrace, we were testing just by trying out the application ourselves and getting a feeling for the performance. That's how it very often would go. You would start up an application and it was judged by the feeling of the person who was using it at that moment in time. That, of course, is not representative of what the actual end-user feeling would be. We were totally blind. We actually need this to be able to be closer to the customer. To really care about the customer, you need to know what he is doing.
Also, incidents are resolved much faster by using Dynatrace. And that's for front-end, because we actually know what is going on. But it's also for server-side incidents where we can see the correlation. Using this solution our MTTR has been lowered by 25 percent. It's pinpointing the actual errors or the actual database calls, so it goes faster. But, of course, you still have to do it. It still needs to be implemented. It doesn't do the implementation work for you.
Root cause detection, how the infrastructure components interact with each other, helps. We know what is going wrong and where to pinpoint it. Before, we needed to fill a room with all the experts. The back-end expert would say, "I'm not seeing anything on the back-end." And the network expert would say, "I'm not seeing anything on the network." When you see the interaction between the different aspects, it's immediately clear you have to search in your Java development, or you have to search in your database, because all the other ones don't have any impact on the performance. You see it in Dynatrace because all the numbers are there. It really helps with that. It also helps to pinpoint which teams should work on the solution. In addition to the fact that it's speeding up the process of finding your root cause, it's also lowering the number of people who need to pay attention to the problem. It's just a single team that we need to work on it. All the rest can go home.
It has decreased our mean time to identification by 90 percent, meaning it only takes us one-tenth of the time it used to, because it immediately pinpoints where the problem is.
Dynatrace also helps DevOps to focus on continuous delivery and to shift quality issues to pre-production because we are already seeing things in pre-production. We have Dynatrace in our test environment, so we have a lot of extra information there, and DevOps teams can actually work on that information.
Finally, in terms of uptime, it's signaling whenever something is down and you can react to the fact that it is down a lot faster. That improves the uptime. But the tool itself, of course, doesn't do anything for your uptime. It just signals the fact that it's down faster so you can react to it.
The most valuable aspect is the fact that Dynatrace is a correlation tool for all those different layers. It's the correlation from the front-end through to the database. You can see your individual tracks.
One of the aspects that follows from that is the root cause analysis. Because we have these correlations, we can say, "Hey it's going slow on the server side because a database is having connection issues," for example. So the root cause is important, but it's actually based on the correlation between the different layers in your system.
Dynatrace is a single platform. It has all these different tools but they are actually all baked into the OneAgent technology. Within that OneAgent — which is growing quite large, but that's something else — you have the different tool sets. You have threat analysis, memory dumps, Java analysis, the database statements, and so on. It's all included in this OneAgent. So the management is actually quite easy. You have this one tool, and you have server-side and agent-side which are ways of semi-automatically updating it. We don't have to do that much management on it. Even for the quite large environment that we have, the management, itself, is quite limited. It doesn't take a lot of time. It's quite easy.
The solution's ability to assess the severity of anomalies based on the actual impact to users and business KPIs is great. It's exactly what we need. The severity impact is based on the users, the availability, and the impact it has on your business.
We also use the real-user monitoring and we are using the synthetic monitoring in a limited way, for the moment. We are not using session replay. I would like that, but it's still being considered by councils within the company as to whether we are able to use it.
We are using synthetic monitoring to measure the availability of one of our services. It's a very important service and, if it is down, we want business to be notified about this immediately. So we have set up a synthetic monitor, which is measuring the availability of that single service each minute. Whenever there is a problem, an incident will be immediately created and forwarded to the correct person. This synthetic monitoring is just an availability check in HTTP. It's actually a browser which is calling up a page and we are doing some page checks on this page to be sure that it is available. Next to the availability, which the synthetic monitoring gives us, we also measure the performance of this single page, because it's very important for us that this page is fast enough. If the performance of this single page degrades, an incident is also created for the same person, and he can respond to it immediately.
Real-user monitoring is a big part of what we are doing because we are focusing on the actual user experience. I just came from a meeting, 15 minutes ago, where we discussed this issue: a slowdown reported by the users. We didn't see anything on the server side but users are still complaining. We need to see what the users are actually doing. You can do that in debug tools, like Chrome Debugger, to see what your network traffic is and what your page is doing. But you cannot do that in production with your end-users. You cannot request that your end-users open their debug tools and tell you what's going on. That's what Dynatrace offers: insight like the debug tools for your end-user. That's also exactly what we need.
Most of the problems that we can respond to immediately are server problems, but most of the problems that occur, are front-end problems, currently. More and more, performance issues are located on the machine of the end-user, and so you need to have insight into that. A company of our size is obliged to have insight into how its actual users are doing. Otherwise, we're just blind to our user experience.
Dynatrace also provides a really nice representation of your infrastructure. You have all your servers, you have all your services, and you know how they communicate with each other.
While it gives you a good view of all the services that are instrumented by Dynatrace — which is good, of course, and that's what it can do — in our case, our infrastructure is a lot bigger than the part that is instrumented by Dynatrace only. So we only see a small part of the infrastructure. There are a number of components which are not instrumentable, like the F5 firewalls, switches, etc. So it gives a good overview of your server infrastructure. That's great, we need that. But it's lacking a bit of network segmentation and switches. So it's not a representation of your entire infrastructure. Not every component is there.
The solution's ability to assess the severity of anomalies based on the actual impact to users and business KPIs is great. In my opinion, it could be extended even more. I would like it to be more configurable for the end-user. It would be nice to have more business rules applicable to the severity. It's already very good as it is now. It is based on the impact on your front-end users. But it would be nice if we could configure it a bit more.
Another area for improvement is that I would like the alerting to be set up a little bit more easily. Currently, it takes a lot of work to add alerting, especially if you have a large environment, and I consider our environment to be quite large. The alerting takes a lot of administration. It could be a lot easier. It would not be that complicated to build in, but it would take some time.
I would also like the visual representation of the graphs to be improved. We have control of the actual measures which are in the graphs, but we are not able to control how the axes are represented or the thresholds are represented. I do know that they are working on that.
I have been using the Dynatrace AppMon tool for six years and we changed to the new Dynatrace tool almost three years ago.
We haven't had any issues with the stability of Dynatrace, and it's been running for a long time. We use the Managed environment, so it's an on-prem service, but it's quite stable. We are doing the updates pretty regularly. They come in every month but we are doing them every two or three months. First we do them in the test phase and then in the production phase. But we have not experienced any downtime ever.
For us, Dynatrace is scalable and we haven't seen any issues with that. We did need to install a larger server, but that's because we have a managed environment. You don't have that problem if you go with the SaaS environment. We don't see any negative impact on the scale of our products, and we are already quite large. It's quite scalable.
In terms of the cloud-native environments we have scaled Dynatrace to, we are using Dynatrace on an OpenShift platform, which is a Docker Kubernetes implementation from Red Hat. We have Azure for our CRM system, which Dynatrace monitors, but we are not measuring the individual pods in there as it is not a PaaS; it's a SaaS solution of course.
As for the users of the solution, we make a distinction between the users who are deploying stuff and those who are managing the Dynatrace stuff. The latter would be my team, the APM team, and we are four people. The four people are installing the Dynatrace agents, making sure the servers are alright, and making sure the management of the Dynatrace system itself is okay.
The users of the tool are the users of the different business cases. That includes development and business. There are about 500 individual users making use of the different dashboards and abilities within Dynatrace. But we see that number of users, 500, as a bit small. We want to extend that to over 1,000 in near future. But that will take some advertising inside the company.
I use Dynatrace technical support on a daily basis. They have a live chat within the tool and that comes for free with the tool itself. All 500 of our users are able to use this chat functionality. I'm using it very frequently, especially when I need to find out where features or functionalities are located within the tool. They can immediately help you with first-line support for the easy questions and that saves you a lot of time. You just chat and say, "Hey, I want to see where this setting can be activated," and they say, "Just click this button and you will be there."
For the more complex questions, you start with tickets and they will solve them. That takes a little bit longer, depending on how complex your question is.
But that first-line support is really a very easy way to interact with these people, and you get more out of the tool, faster.
We purchased the Dynatrace product because we had some issues with our direct channels, our customer-facing applications. There were complaints from the customer side and we couldn't find the solution.
There were also a number of our most important applications that needed more monitoring. We had a lot of monitoring capabilities on the server side and on the database side, but the correlation between all these monitoring tools was not that easy. When they came up with a problem they would say, "Hey, it's not the mainframe, it's not the database, it's not the network." But what was it? That was still hard to find out. And we were missing some monitoring on the front-end. The user experience monitoring was lacking. We investigated a number of products and Dynatrace came out as the best.
We kind of grew into Dynatrace. Our initial scope was quite small, so it was not that complex. Currently, our scope is a lot broader, but it is not complex for us because we have been working with the tool for such a long time. Overall, it's quite straightforward. If you're starting with this product from scratch and you have to find out everything, it can take some time to learn the product. But it's quite straightforward.
We started with the AppMon tool, which was the predecessor to the current tool. Implementing that went quite fast because it was a very small scope. When we changed to the Dynatrace Managed it took us half a year. And that's not including the contract negotiations. That was for the actual implementation: Finding out all business cases and all the use cases that we had, transforming them into the new tool, and launching it live for a big part of our company. That took half a year.
We hired some external experts from a company in Belgium, which is called Realdolmen. They really helped us in the implementation. They had experience in implementing Dynatrace for other companies already, so that really helped. And I would advise that approach. If you're doing it all by yourself, you are focusing on what your problems are, while if you are adding an external person to it, who is also an expert in the product itself, he will give you insights into how the product can benefit you in ways you couldn't have imagined.
The issue of whether Dynatracec has saved us money through consolidation of tools is something we are working on. There are a number of things that we are replacing now by things that are already present in Dynatrace. If you currently have a lot of different tools, it will save you money. But Dynatrace is not the cheapest tool. Money-saving should not be your first concern if you buy Dynatrace.
It depends on your business case, but as soon as you are at a reasonable size and you have different channels to connect within your company — mobile and web and so on — you need to have a view into your infrastructure and that's where Dynatrace provides real benefits. It's not for a simple company. It's not for the bakery store around the corner. But as soon as you hit a reasonable size, it gives enough added value and it's hard to imagine not having it or something comparable.
"Reasonable size" depends a bit on your industry. But it is connected with the number of customers you have. We have about 25,000 concurrent customers, at a given moment in time. As soon as you have more than 1,000 concurrent customers, you need this tool to have enough analysis power. It gives you power for tracking the individual user and it gives you the power to aggregate all the data, to see an overview of how your users are doing. This combination really gives you a lot of benefits.
It is quite costly. Dynatrace was the most expensive, compared to the other products we looked at. But it was also a lot better. If you want value for your money, Dynatrace is the way to go.
In my opinion, the product is extremely good and comparable. We did compare it to AppDynamics and New Relic and we saw that Dynatrace is actually the best product there is. If you are looking for the best, Dynatrace will be your product.
The biggest lesson that I have learned from Dynatrace is that application performance monitoring is very complex, but the easiest part of it is the technical aspect. The more complex thing is all the internal company politics around it. We see a lot of data and if you are targeting some people and say, "Hey, your data bridge is going slowly," they will respond to it very defensively. If they have their own monitoring tools, they can say, "Oh no, my database is going very fast. See my screen is green." But we have the insights. It's all data, and gathering the data is the technical aspect. That's easy. But then convincing people and getting people to agree on what is obvious data is far more complex than the technical aspects.
The way to overcome that is talking. Communication is key.
I'm a little bit skeptical about the self-healing. I have heard a lot about it. I have gone through some Dynatrace instances where they have this self-healing prophecy. I think it's difficult to do self-healing. We are not using it in our company. There is a limited range of problems that you can address with it. It's only if you definitely know that this solution will work for this problem. But problems are always different, every time. And if you have specific knowledge that something will work if a particular problem arises, most of the time you can just avoid having the problem. So I'm a little bit skeptical. We are also not using it because we have a lot of governance on our production environment. We cannot immediately change something in production.
We are using dynamic microservices within a Kubernetes environment, but the self-healing is a little bit baked into these microservices. It's a Docker Kubernetes thing, where you have control over how many containers or pods you want to spin up. So you don't need an extra self-healing tool on top of that.
In terms of integrating Dynatrace with our CI/CD and ITSM tools, we are working on both of those directions, but we are not there yet. We have an integration with our ITSM tool in the sense that we are registering incidents from Dynatrace in our ServiceNow. But we are not monitoring it as a component management system.
We are not doing as much as I would want to for these Quality Gates. That can be improved in our company. Dynatrace could help with that, but I would focus on something else like Keptn, or something else that integrates with Dynatrace, to provide that additional functionality. Keptn would be more suitable for that, than the Dynatrace tool itself, but they are closely linked together. For us, that aspect is a work-in-progress.
I would rate Dynatrace a nine out of 10, because it has really added value to my daily business and what I have to do in performance analysis. It can be improved, and I hope it will be improved and updates will be coming. But it's still a very good tool and it's better than other tools that I have seen.
It's used in two major use cases:
It does full stack monitoring for internal operations, problem diagnostics, APM use cases, and performance management for our customers.
We have multiple instances of Dynatrace running, where about half of them are running in our data centers and the other half are running in the public cloud. Therefore, it's a hybrid deployment. We use a mixture of cloud providers, including AWS, Microsoft Azure (running Kubernetes), and Google Cloud Platform.
We have traditional deployments on VMware virtual machines as well as running stuff in the cloud. We have a couple hundred Kubernetes clusters monitoring using Dynatrace. Dynatrace's functionality in this area is unmatched combined with its full stack visibility, ease of deployment, and completely dynamic changes. The container environments are also dynamic since you have microservices spinning up and down as you go. I have never seen another tool doing this with the same reliability.
Dynatrace has improved our organization through operational support. We also have a large services organization which directly works with customers, and sometimes you run into situations where customers ask how they can improve their applications. Traditionally, these service teams would go for assessments. Eventually, they would even go onsite and through performance workshops with them to find some low hanging fruits that could address, and this was very tedious work. By introducing Dynatrace, you suddenly have real-time data. Then, the process of doing performance reviews switches from workshops or a defined time frame analysis (and then taking actions) to a more continuous approach where you constantly have Dynatrace performance data of the landscape.
Service engineers save a lot of time because they can just go in look at the data and share it with the customer, who has the same view, and say, "Here's an improvement which can be immediately implemented." It's not like a collection of big, multiple findings that are consolidated into one results presentation, then the customer needs to do something. It's more like a continuous performance analysis and improvement process, which is more efficient than those workshops approaches. That's one of the biggest of the advantages that our services team sees because it helps DevOps to focus on continuous delivery and shift quality issues to pre-production.
Dynatrace is tightly integrated with ITSM. It's integrated with ServiceNow, which our support team is using.
We provide a platform, then the customer ships the code and deploys it. Therefore, we rely on testing by the customer, and sometimes, they miss something and it breaks. Then, it doesn't work as expected so we have to step in, and say, "Yes, your site is down," or "It's not functioning properly." We do the analysis because typically the customer says, "Okay, it's not us. It must be you as the service provider." This is where we gain a lot of efficiency. The support team is the first line of defense there. They get the information to determine if they are able to quickly pinpoint the problem. E.g., the customer deployed, then two hours later, issues were occurring. This is when you don't want to waste time. Our support engineers need the visibility so they can immediately be able to communicate to the customer, saying, "Yes, it's on our side," or "It's on your side." If it's on the customer's side, they can let them know exactly where they need to go. This is where we gain most of the time.
It helps our operations that the solution uses a single agent for automated deployment and discovery. If you think about all the work in the past where we had different agents, tools, or scripts deployed to monitor specific aspects of an environment and different tools, then having one agent definitely helps. For example, for our rollout, when we migrated all the different tools to Dynatrace, we did this over the weekend. We installed the agent, then just watched the data and findings coming in, which was a huge benefit. We installed one thing an it discovers everything.
I suppose the solution has decreased time to market for our individual customers with new innovations/capabilities. Dynatrace helps them gain better insights, allowing them to do another deployment faster.
It has auto detection of almost everything. The full stack capabilities to get one agent deployed allows you not to worry about anything else because the agent detects everything. This is in combination with the AI so you don't need to worry about any baselines or setting up any thresholds. This is all done automatically, which brings us the biggest benefit.
Configuration as code integrating through APIs is really important when automating at scale. If you think about the tens of thousands of hosts that you deploy to, then APIs are key when automating deployments, the management of those instances, and configuration as well as integrating with other systems without sophisticated or far reaching APIs.
Dynatrace easily integrates with our infrastructure or applications, then reliably triggers self-healing actions or remediation actions. This is something that we really love to use because it definitely removes a lot of human interaction. You just let the machine to do the job and can trust it, and that's the most important. I have seen systems where the users were very reluctant to trust the system to take actions where typically a human would do the job manually. Dynatrace considers all the information that it gathers, then triggers self-healing actions which are quite reliable. It doesn't need a lot of human adjustment to make it work.
We use real-user monitoring a lot to get insight into end users and our customers, e.g., customer behavior.
While the integrations are great, sometimes our customers are not as far as long in Dynatrace concepts from a technical perspective as they need to be, whether it's a cultural thing and educational thing. Thus, some of our customers are not as advanced as Dynatrace would like them to be. From a technical perspective, all the capabilities are there but the concepts are not yet spread out within the ecosystem to their fullest extent. Therefore, Dynatrace is ahead of its time.
Documentation could be improved. E.g., you don't know how to properly use Dynatrace because documentation is almost lacking behind the features being deployed.
On very large deployment scenarios, the APIs for configuration and configuration management came in slowly. This is something that is good already but could be better.
In the product, I am missing some configuration automation APIs.
The company has been using Dynatrace on different occasions for the past eight years. The current product of Dynatrace has only been out for four years.
We operate services for our customers with pretty high SLAs. We guarantee the systems we run are reliable. We also guarantee uptime. In the past three years, we have run up to 50 updates with Dynatrace and had only one or two issues where the system had to be brought down. There are almost no issues at all with stability. It is rock-solid.
They are improving constantly with every release and adding new stuff. We have updates about every two weeks.
We have about 2,500 people using it.
We currently manage seven Dynatrace clusters with several thousand Dynatrace tenants, then in total almost 30,000 hosts are monitored with Dynatrace. We're not reaching the limits of Dynatrace's scalability. This is probably one of the largest deployments, but we have not seen any limitations so far.
We want to leverage even more services:
There is a lot of data in Dynatrace at the moment that we do not fully utilize.
The technical support is great. We have a pretty good contract with Dynatrace for contacting support. They are pretty responsive and very knowledgeable. You get a DevOps engineer from Dynatrace jumping on immediately with very high expertise. You don't get the typical Level 1 automated standard reply: "Yes, we will take care of it," but then you have to ping back.
We came from a former product of Dynatrace, which was called AppMon, and not really sold anymore. Though, there are customers who still use it out there. We used it for the traditional APM scenario, then migrated to Dynatrace to extend the visibility for hybrid cloud deployment.
We had been using a mixture of Opsview, Splunk, SolarWinds, and other tools. We switched because of the complexity of managing all these tools. It became unmaintainable. E.g., historically, people would write scripts for Nagios Opsview, then maintain them. If we lost the people who had been maintaining those scripts, then nobody knew how the checks worked for those custom scripts. Also, the maintenance overhead was pretty high.
From the perspective of the end users using different monitoring solutions, you had different teams who had to go to different tools and contend with data in one tool not being exactly the same data as another tool. While the overlap between tools was there, the complexity in accessing those tools and knowing how to use those tools became a big organizational and maintenance overhead that we decided to pull them all into one tool to harmonize it. We wanted one tool where the interface and data are the same regardless of whatever you're monitoring.
The initial setup was straightforward. We looked into Dynatrace and were able to roll it out to 12,000 hosts within four weeks.
From the Managed version, you can have it installed and up and running in less than an hour. This is on the condition that you have the hardware to install it on and access to the systems/services that you want to monitor.
Initially, some people were skeptical about the one agent really working, so we did test it. Now, we have had so many good experiences that when we deploy, build new services, or spin up new instances, Dynatrace is one of the first things that is always there. We don't even even test the agents anymore. We completely rely on this mature product that is solid and stable when we deploy staging, development, QA environments, or playgrounds. There is no deployment without Dynatrace agents.
We deployed Dynatrace ourselves as we have a lot experience working with it. Deploying Dynatrace depends on the environments that you run it on. Since that was all orchestrated with things like Puppet, Chef and Ansible for us, it just was a matter writing a bit of automation code that it wasn't already in place. One person was needed to do this properly, and it is not that hard of work because it applies to almost every environment that we deploy. For new services that we provide, it's done within the development teams writing those services. Therefore, there is no dedicated Dynatrace team responsible for integrating Dynatrace with services.
There is almost an API for everything. If you run it Managed, this means you have to administer Dynatrace's installation yourself. You run it and take care of some prerequisites, like sizing. Any system updates, back fixes, or upgrades to the whole cluster have almost zero maintenance. All you need to do is confirm it or let Dynatrace update itself. In the past three years, we had almost 50 updates or installations where we didn't even need to touch anything. We just had one or two occasions where an update broke functionality, and those were fixed with next update and within hours. It's almost self-maintaining.
We do have a dedicated staff for maintenance, but this team is not spending a lot of time on actually managing Dynatrace. They do the integrations of Dynatrace and other tools as well as development of custom integrations and configurations. This team is also responsible for the infrastructure and ensuring the machines Dynatrace runs on are scaled or adjusted properly. However, this is minor effort for them. We have a dedicated team of 20 to 30 SRE engineers and their responsibility is not only to Dynatrace. They are responsible for the whole infrastructure and surrounding tools.
As we use it internally, our internal operations have gained a lot more efficiency. The time to resolution and triage problems in different environments has been reduced by 50 percent, if not more. When Dynatrace raises a problem, the team does not need to bring together experts from other teams to look at the problem, log files, etc. You almost have Dynatrace training our support engineers because it's so easy to pinpoint the root cause of problems.
The solution has decreased our mean time to identification by approximately 50 percent.
There has been a positive impact on the instances run for our customers. Overall, uptime got better because we became faster at fixing the problems causing downtime.
The solution has saved us money through the consolidation of tools. With a hybrid landscape, we had multiple tools. When we consolidated, we removed four or five other monitoring tools with one. For the last ROI calculation that I did, Dynatrace was saving us up to $500,000 per year.
In addition, our speed is up 40 to 50 percent. Therefore, our human cost and licensing savings together are one to two million.
We are a very big customer. We obviously have a special price point.
If there are no corporate requirements to run Dynatrace Managed (operating it yourself), I would definitely go for the size option. For small and medium-sized companies, the size option is probably the cheapest one. You don't need to look into operating it. You don't need to run hardware. It is pay as you go.
We looked into what can Dynatrace could actually replace. If the price point is high, think about the impact it would have to the entire organization to constantly replace monitoring tools. If implemented correctly, then it has a lot of saving potentials for the organization. That is something that should go into any ROI calculation.
We looked at the other big player in this space: New Relic and AppDynamics. Looking at the cloud, full stack capabilities, ease of deployment, and scalability that Dynatrace has, they definitely stood out in comparison. The full stack story was pretty compelling, where you have one agent deployed and it provides everything.
Trust what it's doing. Don't question what it's doing. If you don't understand it yet, take the time to try to understand it. Do not implement or force the old ways of monitoring onto a completely different approach, like Dynatrace. That's definitely that the biggest lesson a lot of people in our organization had to go through.
Be curious and embrace the different approach. It is definitely worth it. The different approach that it does is a good one. It's different but it's something that actually works. Those guys know what they have built and what they are doing.
It is partly integrated with CI/CD. We are operating a platform with our applications, but our customers are responsible for testing and CI/CD deployed into our environments. Internally, some of our teams use it. The majority of our CI/CD deployment is our customers' responsibility, and while we do provide them Dynatrace for CI/CD, we do not control how they integrate it.
We are in the process of rolling out synthetic monitoring at scale to replace other tools.
We are not yet using session replay, which is mostly due to data compliance restrictions. We have very hard data privacy protections. We do have customers who are highly interested in using the feature, but we are not using it at the moment.
Overall, I would give the solution a clear 10 (out of 10).
Our primary use cases are operational awareness, health of the systems, and impact on users. Other use cases include proactive performance management, system checkouts (as we investigate the ability to manage configuration and integration to the CMDB), some usage of it from a product perspective in terms of application usage, and I use it to manage and improve the user experience by understanding user behaviors.
We are in both Azure and AWS. We have both on-premise and cloud Kubernetes environments that we're running in. In fact, we have been using less efficient deployment methodologies. We haven't encountered any limitations in scaling to cloud-native environments.
We have only used version 1.192 of the Dynatrace product. We have not used any previous versions.
It has improved our critical incident response, exposing critical issues impacting the environment and our ability to respond to those events prior to client impact as well as resolving those events more quickly. We have use cases where we have studied a 70 percent improvement for response times in an occurring event as well as future reoccurrences being improved.
The solution's use of a single agent for automated deployment and discovery helps our operations significantly. Oftentimes when you are looking at endpoint management, centralized monitoring teams need access to data across systems. They need to manage agents deployed throughout the organization. Remote polling of data can be helpful, but it's not deep enough, especially for APM capabilities. Having one agent significantly simplifies that functionality in such a way that it enables a very small team to manage a very large environment with very limited overhead. It provides the ability for external teams to manage it because they don't need any deeper knowledge of the application than installing the agent. They have the ability to integrate the agent into deployments and to do the work with very limited overhead.
The automated discovery and analysis helps us to proactively troubleshoot production and pinpoint the underlying root cause. We have had scenarios where we can see end user impact. One of the use cases was where we had an individual system and a cluster of nine for a content management system that was having an issue. Through Dynatrace, we were able to quickly identify the one host that was having a problem, take that out of the active cluster, recycle that application instance, bring it back, and reintroduce it to the cluster in a very efficient manner. Historically, these processes take multiple hours in order to diagnose and identify the instance, then do the work. With Dynatrace, we are able to do the work in less than 20 minutes from when it first occurred to issue resolution. Thus, there have been scenarios where you can quickly identify infrastructure issues and back-end services.
Out-of-the-box, it's the best product that I've seen. Its ability to associate application impact, as well as root cause from an infrastructure standpoint, is by far ahead of anything that I have seen due to its ability to associate infrastructure anomalies to applications. We are still on our journey of identifying the right business KPIs to see how we can associate this data.
Dynatrace is doing an excellent job of giving us 360-degree visibility of the user experience across channels in most technologies. We are working with Dynatrace to expose the full transparency to the mainframe, as we have transactions that call from the cloud onto the mainframe and back out to other services. This is a critical visibility that isn't there yet. Otherwise, with a lot of the cloud and historical systems, we do see a lot of transparency of transaction trace across the environment.
These are probably the most key, because it gets into the traceability from tracing transactions of the end user all the way through the back-end systems. We are still working through the mainframe integration, but the scenarios where we can integrate through the mainframe are very useful.
We can see issues that occur, sometimes before the clients do. Before we have client (or end user) calls for issues, we are able to start troubleshooting and even resolve those issues. We can quickly identify the root cause and impact of the issues as they occur, and this is very helpful for providing the best client experience.
We have found the self-management of the management cluster and Dynatrace processes to be highly reliable. There have been minimal issues with managing the infrastructure.
We've targeted deployment of the real-user monitoring to the most critical applications in the company to understand if there's something that's happening in the environment and the user impact. This is to be able to understand the blast radius of issues, helping us understand if an issue is impacting one app or multiple applications. We can then quickly diagnose where the common event is (the root cause), resolve it, and then leverage the product to validate healthy user traffic after completion by seeing transactions be processed again.
From a synthetic standpoint, we use the synthetics in two ways:
It has been very easy to deploy and obtain basic information.
It's very good from a problem troubleshooting perspective.
I find the value from the out-of-the-box features to be extremely valuable. However, there will be gaps and challenges as you go into a much broader set of infrastructure technologies to consume that necessary information. This will be a challenge for the company. The things that they need to focus on is the ease of integrating external data sources, which can then also contribute to the AI. There is a ton of value gotten out-of-the-box, but moving to the next steps will be an interesting journey. I know this is something they are focused on now. When bringing in other telemetry, whether it be network devices, databases, or other third-party products that all integrate into a larger ecosystem, there will also be a lot of successes, but there will also be some challenges on this journey.
There is some complexity in the alarm processing logic within the product between the alert policies and problem notifications.
Expand the user session query data to be inclusive and enable that for the application or other telemetry within the system. Currently, in order to analyze the data outside of dashboards, it requires exporting to other reporting systems. If you want to do higher level reporting, then this may make sense. However, there is a desire to be able to do some of that analysis within the product.
There continues to be some opportunity to expose the infrastructure from a broader reporting standpoint. Overall, the opportunity is in the reporting capability and the ability to more flexibly expose or pivot the data for deeper analysis. Oftentimes, the solution is good at looking narrowly at information, but when you want to broaden that perspective, that's where the challenges come in. At this point, it requires the export of data to external systems to do this.
Adoption lagged primarily due to:
About two years.
At this point, we have about 1700 host units. We're monitoring 2000 to 3000 systems. We have 300 to 500 users a month using the systems with approximately 700 users overall.
Their Tier 0 is better than most companies that I have ever worked with. Normally, I'll get useful information even at that initial level/Tier 0.
The in-app chat is extremely helpful. It helps not only with the ability for me to troubleshoot, but the ability for the rest of the organization to ask how-to questions. We have hundreds of those chats across the organization per month which are leveraged by end users.
Everything else is as expected when working through engineering and our product specialists, who have been helpful.
The initial setup and implementation are almost too easy. With real-user monitoring and all the application monitoring, you are introducing change into the environment. It is so easy to set up, configure, and implement that you can get way ahead of your organization technically from where they are from a usability standpoint. We have run into virtually no technical limitations in implementing the product. It has purely been from the ability to get users to adapt, understand, and leverage the value of the product.
We implemented and installed the Dynatrace platform (and everything) within a couple of days. We deployed the product in certain environments within overnight of instrumentation. Onboarding of teams and the training required, that took months. Even though we were able to technically implement the product from non-production into production within a month of deploying everything, having it there, and instrumented. It took us another eight to nine months to onboard individual teams into adopting and leveraging the product. From there, the rolling out is really limited more by organizational change, communication, and facilitating training with teams and their technical capabilities. Key teams have adopted the product and used it very quickly. Therefore, we are seeing value within four weeks of deployment from our centralized critical incident teams, but the product adoption from application and development teams has lagged.
If you are implementing Dynatrace, the first thing is to not underestimate your users and their experience, providing them personal service to onboard and consume the information, then leverage the product on the front-end. Technically the product makes it so easy to implement and deploy, this makes it difficult to stay in front of the rest of the organization when adopting the product. You need to ensure the data starts presenting itself before they are ready and able to consume it. You need to focus that into your implementation.
The solution has decreased both our MTTI and MTTR.
In 2018, we were having on average one issue per day. It is one of the reasons that we purchased the product in 2018. Last year, we significantly drilled those numbers down in outage time by 70 to 80 percent, as an organization. While Dynatrace is part of driving that avoidance as well as reduced outage time, it's impossible for us to have a direct correlation of its direct impact because there are so many other factors at play in an organization. I had to change management processes and everything else that could also influence that. However, we know that it was part of that increased uptime to where we've decided to invest significantly more in the product.
It's understandable to do a smaller scale initial evaluation. However, as you identify the product value, don't hesitant in your scope and scale to maximize the initial investment and your opportunity to do a bulk investment of the product.
We have other competitive products. The automation instrument will be extremely valuable as we look to consolidate our solution set. The insight to quickly gain information is interesting and good information that we can use. There will be a challenge internally with our teams since application teams were never exposed to infrastructure information and infrastructure teams have never been exposed to application nor end user information. Organizationally, we have to change where people are now going to see this insight and figure out how to leverage it for good, which will be helpful. It will be a game changer in terms of how we can identify and respond to events in the organization from the point of view of data and analysis, as opposed to tribal knowledge and fear.
Dynatrace was initially brought in to eliminate one competitive APM product. We are now on to eliminating the second, and we'll be consolidating all APM on the Dynatrace platform. We are also in the process of consolidating other infrastructure monitoring products on the platform. We expect there will be a small incremental investment from a purely licensing standpoint to consolidate the products, but we expect realization of a significant amount of benefit from the capabilities it provides from root cause analysis, impact analysis, transaction trace observability in the environment, the reduced administrative costs of disparate products, and the ability to integrate data. However, a lot of these were not measured previously because we had a lot of disparate tools across disparate teams managing things. Therefore, we can't measure the savings but we expect it will be significant.
We have CA APM Introscope, New Relic, and AppDynamics. We are users of all three of these products, though we are probably using AppDynamics the least. We have almost completely migrated away from Broadcom and are starting the replacement of New Relic.
Holistically, Dynatrace's traceability starts from the user endpoint, meaning the ability to trace a transaction from a user session all the way through other technologies. We've had more comprehensive traces than with other products. Other products do not offer an easy interface to see the trace of the user session in a comprehensive way. Dynatrace offers the ability to go from a mobile, microservices, or mainframe and be able to trace across all those platforms. It also has the ability to associate or automatically correlate user transactions to applications, then into the underlying infrastructure components. Another Dynatrace benefit is the whole function of the AI as well as bringing in other external data sources. E.g., we are looking at things like a DataPower and F5 data integrations, but also incorporating those into the trace. Finally, there is support of legacy technologies, because it really gets into traceability, AI, and the supportive legacy. Mainframe technologies are the big positive differentiators and kind of come to a conclusive root cause analysis.
CA APM Introscope and New Relic have simpler interfaces to consume data. With Dynatrace, you need to develop plugins to obtain easier API interfaces for pushing data into other products. This is a little easier with the other products. The New Relic Insights product is a stronger reporting feature than what Dynatrace provides.
There are also other products that we are looking at eliminating in other product suites, such as Broadcom UIM, Microsoft SCOM, and Zabbix. We have a lot open source solutions where we're looking to roll out infrastructure, then consolidate and centralized data. The primary function and capabilities gets into mobile to mainframe traceability in order to simplify or expedite impact and root cause analysis processes for the teams. The solution also has the ability to support our modern technologies running in AWS and Kubernetes cluster microservices as well as traceability all the way through the mainframe.
We have integrated our notification systems through PagerDuty, Slack, and our auto ticketing app. This is to generate incident records. The integrations with PagerDuty and Slack are effective. We're in the process of migrating some tools to ServiceNow. Thus, we are in the process of doing synchronization of both the events while also evaluating the CMDB integration with ServiceNow. There are some recent capabilities that make this look more attractive to automate discovery and relationship building that we're looking forward to, but we have not yet implemented. The integration to ServiceNow will be good.
The desire is to have Dynatrace help DevOps focus on continuous delivery and shift quality issues to pre-production. We are not there yet. The vision is there and it makes sense with the information that we see, but we have not had the opportunity. Even though we've been using the product now for two years, we're only now just starting an effort to roll the product out across the enterprise and replace competitive products for application infrastructure monitoring. We'll then have the opportunity for that full CI/CD integration or NoOps opportunity.
We will be rolling out to some highly dense environments in the near future. We haven't run into any performance issues yet. The only issue that we ran into previously is with the automated instrumentation of the product. We accidentally disabled the competitive products that teams were using as we were evaluating Dynatrace. You can get in front of yourself in rollout.
We don't have the solution’s self-healing functionality integrated into the automation product. Dynatrace doesn't have the self-healing capability of restarting services. Therefore, from a monitored application perspective, we haven't enjoyed that capability yet.
We are in the process of testing some parts of the session replay. We see value there and are working through understanding the auditory or compliance impacts to leverage this feature.
Based on my experience and history of the products, I would rate it at least a nine (out of 10). It's been far superior to other products in its capabilities and comprehensiveness, especially across both cloud and legacy technologies, such as older technologies (like mainframes and server-based monolithic applications).
We are using it to monitor our e-commerce applications and the full stack that our e-commerce applications run on. That includes both our Rack Room Shoes domain and our Off Broadway Shoes domain. We use it to monitor the overall health of the entire stack, from the hardware all the way to the user interface. And more specifically, we use it to monitor the real user's experience on the front-end.
What Dynatrace has really allowed our team to do is focus more on innovation, rather than on monitoring and bug-squashing. Now that we have a tool like Dynatrace, we can continue to do forward-thinking projects while Dynatrace is doing the monitoring and rooting out the root causes. We're spending a lot less time trying to find out what the problem is, versus letting Dynatrace pinpoint where the problem is. We can then validate and remediate much quicker. That's the impact it's had on our business.
The automated discovery and analysis helps us to proactively troubleshoot production and pinpoint underlying root cause. We recently had some issues with database connections. Our database team was scratching their heads, not really knowing where to look. What we were able to do with Dynatrace, because we had some of the Oracle Insights tools built into the database, was to provide, down to the SQL statement, what queries were taking up the most resources on that machine. We provided that to the database team and that gave them a head-start in being able to refactor the data so it was quicker to query. That really helped us speed up the user experience for that specific issue.
Dynatrace helps DevOps to focus on continuous delivery and to shift quality issues to pre-production. We are just now starting to use it in that way. When we first launched Dynatrace, we only had monitoring in our production environment. At that point we were using it as an up-front, first-alert tool for any issues that were happening. Now what we're doing is instrumenting our lower environments with Dynatrace so that it will allow us to monitor our load-testing in those environments, to find out where our breaking points are. So it does allow us to push out products that are much more stable and much less buggy because we're able to find out where our breaking points are in the lower environments. What this is going to do is allow us to do is push out, at a faster rate, more solid, less buggy releases and customer features, and allow us to continue to innovate on the next idea. We're just starting that journey. We just got fully instrumented in our lower environments in the last couple of weeks.
In terms of 360-degree visibility into the user experience across channels, we're only monitoring our digital channels right now, specifically our e-commerce channels. But we do have ways, even within the channel, to dissect by the source they came from. Did a given customer come from a digital ad? Did they come from an email? Did they come to us direct? It does allow us to segment our customers and see how each segment of customer performs as well. This is important for us because we want to make sure that we're not driving specific segments of customers into a bad-performing experience or to a slow response time. It also allows us to adequately determine where to spend our marketing dollars.
Another benefit is that it has definitely decreased our mean time to identification, with the solution and the Davis AI engine bringing the most probable root cause to the top. And within that, it gives us the ability to drill down into the specific issue or query or line of code that is the issue. So it has saved us a lot of time — I would estimate it has saved us 10 hours a week — in remediating issues and trying to find the root cause.
It has also improved uptime, indirectly. Because it gives us alerts early, we're able to mitigate issues before they're actually bigger issues.
The alerting systems are definitely the most valuable feature. The AI engine, "Davis," has proved to be a game-changer for us, as it helps to alert us when there are anomalies found in our applications or in their performance. We find that very helpful. There's still a human element to the self-healing capabilities. I wish I could say, "Oh, it's magic. You just plug it in and it fixes all your problems." I wouldn't say that, but what I would say is that the Davis engine gives us that immediate insight and allows us to cater to our solution so that the next time that problem arises it can mitigate it without a lot of human involvement.
Dynatrace's ability to assess the severity of anomalies, based on the actual impact to users and business KPIs, is really good, out-of-the-box. But it does an even better job when, again, we as humans give more instruction and provide more custom metrics that we're trying to monitor that are key to our business. And then, letting the Davis engine find those anomalies and push them to the top, especially as they relate to business impact, is very valuable to us.
We find the solution's ability to provide the root cause of our major issues, down to the line of code that might be problematic, to be valuable.
And we get a lot of value out of the Session Replay feature that allows us to capture up to 100 percent of our customers' real user experiences. That's helped us a lot in being able to find obscure bugs or make fixes to our applications.
We also use real-user monitoring and Synthetic Monitoring functionalities. We use real-user monitoring for load times, speed index, and overall application index. And we use Synthetic Monitors to make sure that even certain outside, third-party services are available to us at all times. In certain cases, we have been reliant on a third-party service, and our Dynatrace tool has let us know that that service isn't available. We were able to remove that service from our website and reach out to the service provider to find out why it wasn't available.
We also find it to be very easy to use, even for some of our business users. Most of the folks who use the Dynatrace tool do tend to be in the technical field, but use is spread across both the business side, what we call our omni-channel group, as well as our IT group. They all use it for different purposes. I'm beginning to use it on the business side to show the impact that performance has on revenue risk. I can then go back and show that when we have bad performance it affects revenue. And I can put a dollar amount on that. So the user interface is very easy to use, even for the business user.
Dynatrace continues to innovate, and that's especially true in the last couple of years. We have continued to provide our feedback, but the one area that we get value out of now, where we would love to see additional features, is the Session Replay. The ability to see how one individual uses a particular feature is great. But what we'd really like to be able to see is how a large group of people uses a particular feature. I believe Dynatrace has some things on its roadmap to add to Session Replay that would allow us those kinds of insights as well.
We started using Dynatrace in September of 2017. At that time it was an older product called AppMon. But we quickly upgraded to the current Dynatrace platform the following year. We've been using the SaaS platform ever since.
It's been very stable. We've had very little downtime. In the last four years there may have been one outage. Overall, it's been extremely stable. Many times, Dynatrace is our first alert that we have issues with other platforms.
It's extremely scalable. We're one of the small players. We're running with about 70 agents right now. We've been at Dynatrace's conferences and have heard of customers who can deploy 5,000 agents over a weekend and have no issues at all. For our small spec-of-sand space, it's extremely scalable.
We are hosted on Google cloud. That's where all of our VMs are currently set up. Our database is there, our tax server is there. All of our application and web servers are there, and Dynatrace is monitoring all of that for us. We haven't encountered any limitations at all in scaling to our cloud-native environment. We can spin up new auxiliary servers in a matter of minutes and have Dynatrace agents running on them within 15 minutes. We're starting to play a little bit with migrating a version of our application into a Kubernetes deployment and using Dynatrace to monitor the Kubernetes containers as well.
We have plans to increase our usage of Dynatrace. We just recently updated our hosts. We needed to increase the number of host units so that we could put Dynatrace on more servers, and we've already just about used up all of those. So next year, we'll likely have to increase those host units again. And we're going to start using more pieces of Dynatrace that we haven't used before, like management zones and custom metrics.
Technical support has been great. The first line of defense is their chat through the UI, which is really simple. They're super-responsive and usually get back to us within minutes. We have a solutions engineer that we can reach out to as well, and they have been very helpful, even with things like setting up training sessions and screen-sharing sessions to help enable our internal teams to be more productive using the tool.
We were using a tool called New Relic and we were really just using it as a synthetic monitor to make sure the application was up and running, but we really weren't getting a lot of insights. When we decided that we wanted a tool that could give us more insights and that we needed a tool that could give us the ability to monitor more of our customers' behaviors, there just wasn't another tool like Dynatrace that we felt could do things as well as Dynatrace, through a "single pane of glass." We chose Dynatrace over New Relic at the time because New Relic just didn't have any solutions like it.
We haven't found another tool that can help us visualize and understand our infrastructure, and do triage, like Dynatrace. We haven't found one that can give us that full visibility into the entire stack from VM all the way to the UI. That was really the reason we picked Dynatrace. There just wasn't another tool that we felt could do it like Dynatrace.
The fact that the solution uses a single agent for automated deployment and discovery was the second reason that we chose Dynatrace. The ease of deployment, the fact that we could use the one agent and deploy it on the host and suddenly light up all of these metrics, and suddenly light up all of these dashboards with insights that we didn't have before, made it extremely attractive. It required a lot less on our part to try to do instrumentation. Now, as we add more Dynatrace agents to more of our back-end servers, we think we'll gain even more value out of it.
We started with AppMon, which was more of an on-premise version, where we were installing it, although it still was a one-agent. Then we moved to the SaaS solution, and it was very easy for us to migrate from AppMon to the SaaS solution, and it's been extremely easy to instrument new hosts with the agent.
We were up and running within 30 days when we were first engaged with AppMon. When we migrated to the SaaS solution, it maybe took another 30 days and might have even been less. I wasn't involved with that migration, but I worked closely with the guy who was. I don't remember it taking much longer than 30 days to migrate.
We had an implementation strategy. We knew specifically which application we wanted to monitor, and all of the hardware and services and APIs that that application was dependent on. We went in with a strategy to make sure that all of those things were monitored. And now we've progressed that strategy to start monitoring more of our internal back-end systems as well — the systems that support our stores, not just our e-commerce channel — to see if we can't get more value and maybe even realize more cost savings on our brick and mortar side using Dynatrace.
We have definitely seen return on our investment. It has come in the form of being able to produce more stable, less buggy applications and features, and in allowing our team to focus more on innovating new ideas that drive revenue and business, versus maintaining and troubleshooting the existing application.
It hasn't yet saved us money through consolidation of tools, but as we continue to find more value in Dynatrace, it does make us look at other tools and see if we are able to use Dynatrace to consolidate them. We have replaced other application monitoring tools with Dynatrace, but we've not yet consolidated tools.
Whatever your budget is, you can manage Dynatrace and get value out of it, but you need to manage it to what your needs are. That's the one thing we found. We did not budget the right amount to begin with. It has cost us more in the long run than if we would have been able to negotiate it upfront. But we didn't really know what we didn't know until we'd been using Dynatrace for awhile.
Your ability to catch your Session Replay is based on the number of what they call DEM units, digital experience monitoring units. That's where we were short to begin with. There is an additional expense to determining not just the platform subscription but also the number of hosts units that you want to run and the number of DEM units that you need to be able to capture all of the user experiences that you want. In our case, we wanted the ability to capture 100 percent. Maybe in another business someone would only be worried about capturing a sampling of the traffic.
We evaluated New Relic, AppDynamics, AppMon, which was the Dynatrace solution at the time, and we also looked at Rigor.
Dynatrace could do pretty much everything. It wasn't just the real-user monitoring piece of it. It was also the full stack health aspect. The Davis AI engine was probably the biggest differentiator among all of the tools. The Davis AI engine and its ability to surface the root cause was a game-changer.
My advice would be to jump all-in. There doesn't seem to be another tool that can do it like Dynatrace, and from what we've seen the last two times we've gone to their Dynatrace Perform conferences, they are dedicated to innovating and adding features to the platform.
We are not yet using Dynatrace for dynamic microservices within a Kubernetes environment. We are beginning to play in that arena. We're looking at tools that will help us migrate from our current VM architecture to a Kubernetes deployment architecture, to enable us to get more into a no-DevOps type of environment. But today, we're still on a virtual machine deployment architecture.
Similarly, we have not integrated the solution with our CI/CD and/or ITSM tools. That is on our roadmap. As we migrate and transition into a no-DevOps and continuous improvement/continuous deployment operation, we'll begin to use Dynatrace as part of our deployment processes.
The solution hasn't yet decreased our time to market for new innovations or capabilities, but we believe that we will realize that benefit going forward, since we'll be leveraging Dynatrace in our lower environments to find out where breaking points are of new features that we release.
We have half-a-dozen regular users who range from our e-commerce architect to DevOps engineers to front-end software developers. My role as a user is more of a senior-level executive or sponsor role. We also have some IT folks, some database administrators and some CI people, but most of our users are in the IT/technical realm.
We don't have a team dedicated to maintaining the solution. We do have a team responsible for it, though. That is the team that just helped instrument our lower environment with Dynatrace. We've got some shared responsibilities and some deployment instructions that are shared across three different groups. They're from IT, our omnichannel group, which is really our business side, and we leverage a third-party for staff augmentation and they use Dynatrace to help us monitor during our off-hours.
The primary use case is performance, capacity, and availability management along with user experience monitoring of 20 systems on a variety of technology stacks. User experience monitoring and optimisation of system performance and workflow. It has created good visibility on these topics for audit and compliance purposes, supporting adoption of a DevOps culture and practices within the team.
We have FACT, COLLATE, and CODIX iMX technologies as well as in-house developed Java and .NET applications. These are hosted on Windows and Linux OSs and primarily on SQL Server and Oracle RDBMS.
It has created total transparency between technology and business on all aspects of systems and performance as well as being a proxy for network performance through user experience monitoring. This followed a major performance degradation of our primary frontline system, which highlighted inadequacy of infrastructure focus tools, e.g., Nagios and Zabbix. It helped detect and remediate several performance issues on systems on both vendor supplied packages as well as in-house developed systems. It also improved InfraOps and development teams understanding of system behaviour and performance characteristics.
I would like more flexible data export functions and APIs. The end user experience data is very useful to the solutions team to determine actual system usage and misuse. Flexible, easier data APIs would allow us to export the data more easily to other analytics platforms to enable this analysis as well as enable storage of this data for longer term analysis since DynaTrace only holds user data for 35 days.
When we use the Dynatrace API to extract the data it only allows for 5000 records or less, and the data is not sufficiently granular for our needs.
Dynatrace can be configured to continually send user session data to a HTTP Webhook endpoint. Our user session export sends JSON data for all monitored user sessions to the configured HTTP endpoint(postgresql db).
The data is sent in bulk to improve performance, with a flush every few seconds to keep the data rate near real-time.
The data format is one JSON document per line, so we must split the data by line to get valid JSON documents.
We are raising an RFE with DynaTrace to have this data more easily accessible via API
It is very stable with frequent updates and feature expansions.
It is very scalable. Agents limit its own consumption, no longer impacting server hosts.
Our experience was very good. Online help via in-app chat was very helpful. Excellent webinar and online training was provided.
A host of open sourced tools, which could not get beyond basic infrastructure resource monitoring. We needed APM and UEM.
It is easy to get the basics, but more complex when you want more complex metrics and dashboards. E.g., we mapped IP addresses so we knew which corporate campus end users were connecting through it.
We used both. IT Ecology was the vendor. They had excellent knowledge and were able to transfer knowledge to our staff.
It's more expensive than other solutions, but worth it. We use full APM monitoring on our primary systems, but only resource monitoring on lesser systems. We shift licenses around our environment when a deeper dive into lesser systems is required.
We did not really evaluate other options. AppDynamics could do the job, but we had access to an experienced Dynatrace service provider which enabled us to accelerate implementation, rollout, and knowledge transfer.
Anecdotally, it is not as user-friendly as AppDynamics when it comes to configuring dashboards, etc. However, I do not have personal experience with AppDynamics and cannot say for sure.
We use this product for troubleshooting performance issues in our growing set of IoT portal solutions.
This solution has helped us to unearth quite a few issues that were present for a while. Also, it has helped us improve the performance of many of our applications.
The most valuable features are code-level visibility, real user monitoring, and response time statistics.
This solution could be improved with better compatibility with legacy applications.
In our case, we were not able to deeply monitor proprietary solutions such as Windows CE or some BI/ERP applications, which either got crashed or slowed down noticeably.
These kind of issues are bound to happen considering the intrusive nature of the instrumentation process (which is necessary to get good performance insights).
However, some of these technologies would often warrant quite an extensive amount of work to be made compatible, if even feasible.
I am a Dynatrace consultant, and I work with a partner in South Africa.
Once you have it running, Dynatrace will show you a picture of your environment that nobody else would have, except perhaps for the architect. IT environments are inherently complex, and this will help figure out what you've got in the environment.
The most valuable feature is the AI. In the older version, it would highlight errors but you still had to figure out the root cause. With the latest version, the AI engine highlights the root causes automatically.
It would be nice if there were a way that it could be made simpler, given the complexity of the things that we're monitoring. It can get a bit overwhelming. The AI has helped in this regard.
This solution has been very stable.
I am in a small shop so I do not have direct experience with scalability, but I do know that it is one of their design goals. In our company, we have a couple of guys working on huge sites and scaling farms, so my impression is that the scalability is good, or even excellent.
The technical support is good.
The initial setup is pretty straightforward. The agent installs and configures everything automatically. In the older one, you had to manually configure everything. Now, you install the agent and you may have to restart some processes, but then it just starts giving data.
In this solution, they try to take a whole lot of complexity and make it look simple. It is not an easy thing to do.
There is a release every month of new features. They are pretty good at implementing things when you log a feature request or enhancement. The vendor is running DevOps and has a high frequency of releases, so you don't normally have to wait for the next major release, provided that there is enough requirement for it.
Quite a lot of the new design in the new version has been influenced by the new privacy rules in Europe. They've had to restrict a lot of what can be seen, in terms of the user's personal data, which can be seen as a good thing. Generally speaking, they've gone from a very open design where you can see all of the database queries and the data, to a more closed system. You can still find that stuff, but you have to turn on a lot of things and implement them. They are not there by default.
I find this a source of frustration because some of the time, the problems are because of the data. For example, someone put in their name wrong or put in an apostrophe. Without seeing the data, you don't know and can't figure out what is wrong. You have to figure it out by looking at it. This is a GDPR thing, however, and it is necessary for compliance. Companies have to decide while consulting with their customers, how much people are allowed to see. Then it can be configured.
My advice for someone who is implementing this solution is to take some time to plan out your operational environment in advance. Try to maintain consistency in naming, because I think that you can get additional value through this planning. You can roll out ad-hoc and it will be fine, but if you take some time to name things then you can get a better picture of your environment.
This is a very good solution, but nothing is perfect.
I would rate this solution eight out of ten.
We use this solution for monitoring a monolithic application in ASP Classic 3.0 over IIS server. The database is SQL Server 2016 in an AlwaysOn cluster. Over three IIS instances for the core application, two IIS instances for reporting services, and two instances for the batch process in .NET 4.0.
The major improvement was the ability to find errors immediately and predict future failures, or when resources reach the maximum capacity.
The most valuable feature is Session Record because the developers can reproduce an incomplete issue after it is reported.
This solution needs better support for security and monolithic batch processes.
We use this solution for APM and synthetic monitoring, as well as real user monitoring and other basic host monitoring. We have problems sent to PagerDuty via webhook in Dynatrace. This also goes to our on-call center.
Debugging the worst transactions and other slow stuff is also done with this solution.
This has very much improved our organization. We now have a great solution for doing Application Performance Monitoring and doing synthetic checks.
The most valuable feature is the ability to perform synthetic checks for monitoring sites using click paths. Also, real user monitoring and basic host monitoring are really useful features.
Problem analysis is also great.
The user interface needs improvement. Sometimes it is not really clear how you can get to the right place, where you can find all of the information you need. It is sometimes really difficult to understand. If the user interface were made more intuitive then it would really benefit the product.
In Dynatrace, I really do not see any bugs.
Previously we used SolarWinds but it was really buggy.
Installation of OneAgent is really easy. We have installed this solution in a short time on nearly one hundred hosts.
We perform automated monitoring using the Dynatrace API within our CI/CD jobs, and further, synthetic monitoring and HTTP Checks from different locations.
Dynatrace helped us to get a better understanding of how our services are communicating with each other, as well as better problem detection before something really breaks. We have been able to consolidate different tools, like standard host monitoring and synthetic checks.
The OneAgent with Network Flow detection is really amazing. It is simple to install, where you run just one command and you are finished. It automatically detects which applications are running, and injects the tracing automatically without any adjustment in our deployment.
The configuration options should be better accessible. Sometimes it is hard to find the right setting for what you want to change. In K8s deployments, the configuration of the Active Gateway sometimes changes, and when it's automatically updated the monitoring breaks and you don't know why.
This solution is used to create dashboards for business-related applications.
This solution has helped us with improving the monitoring and RCA.
The most valuable feature is the 360-degree view for monitoring, infrastructure, and application worlds.
This solution needs improvement in terms of automation.
We use this solution for application monitoring across regions. It can give a quick overview of the current health status, end to end, and can quickly point to the root cause if there an issue or problem in one of our applications. It helps our managers to keep track of everything on the dashboard.
Dynatrace is not only used by the IT Operations department but also by sales and customer service. It provides an overview of the service that we offer to the customers. The main benefit is that it helps the development team deploy more quickly and with better quality.
The features that we found most valuable are the monitoring, alerting, and Davis the AI. With this combination of features, the Operations department and developers can sit back and relax. They can concentrate on how to improve the service for the customer and work towards a better overall experience.
We would like the AI to produce more scientific data with less configuration. That will help us, as the customers will focus on integrating all of the IT, without hassle.
We need more options for sharing and exporting reports to other systems and platforms.
This solution is used for:
We have a DevOps team distributed between several offices and we need to keep track of every microservice evolution. Dynatrace helps us to do that, keeping the same dashboards and same links for problems in our production environment, or profiling at staging.
I think one of the most valuable features is the Dynatrace API, for both metrics and configuration. With that API, we were able to launch automatic tasks or jobs in response to a specific condition. We also have the option to open a JIRA ticket to keep track of a particular issue.
I think Dynatrace needs improvements with respect to reporting; not just performance, but the business-level reports. The navigation can be improved because when you press the back key, sometimes you lose the time frame. Also, when you are in a problem description and want to leave, it is hard to do.
We use this solution for end-user experience, infrastructure monitoring, analysis of bounce rates, service calls to the database, root cause analysis, and problem management. For end-user analysis, I can monitor where the connections come from, the time, the number of navigated pages, bounce rates, and finally, if the usage was satisfactory or not.
This solution has improved our organization in several ways, including the speed of detecting problems, predictive maintenance, root cause analysis, and alert generation. We are also better able to understand trends with respect to user behavior like time zone connections, and the times when there is less usage of the system by users.
Our organization is too IT oriented. I think with Dynatrace, it has helped to bring value to the business because now we can speak using the same language.
The most valuable features for me are end-user analysis and problem detection. I am responsible for adoption, availability, and performance. In the case of adoption, the number of new users coming to the system is a good metric for management. For problem management, problem detection is a good feature to save time.
I have reported a bug where a CI was not reflected in the dashboard, yet it was detected in the problem management.
On the side of the end user experience, I would suggest adding a new service for analyzing the backtrace of users.
Also, I would like to see an option to export the dashboard to create better reports and avoid copy/paste.
In terms of stability, it is ok and we have had no issues reported so far.
We cannot properly address scalability yet.
Prior to this solution, we used Gomez. It was part of the original solution that was installed. We had many problems with synthetic monitoring because it was down most of the time.
The tools were installed before I joined the company.
There are long term benefits in using the monitoring tool. There is also strategic value added, as is the case of transforming the internal language of the technical teams.
This solution is used to monitor an IoT customer portal. It is hosted on Amazon Web Services. It is using Drupal, Microsoft SQL server, and eWON. By using Dynatrace Managed, we have given recommendations to developers and application owners on coding practices, but have also raised alerts on performance and availability.
By using Dynatrace Managed, we have improved customer satisfaction and have also helped internal teams and subcontractors make the application more stable. Communication between different teams has also been made easier and more fluid.
Dynatrace Managed has given us an in-depth, comprehensive view of every single layer of the entire application stack in a single, easy-to-use point of access. We have had several wins and increased our credibility among all actors involved with the application.
If we can gain more insight into older applications, using not-so-recent technologies, then it would be a plus. Also, having more flexibility to have different user profiles could be beneficial. No real frustration though.
I am a Dynatrace partner and the company where I work is a system integrator. We propose the Dynatrace solution to our customers for monitoring the performance of applications and integrate the solution with the other tools that they have.
Our customers are now able to control their infrastructure with a single tool and can easily check the connections with all components. With the Smartscape and AI, they can see everything and have reduced the number of events.
The ability to monitor a hybrid environment is very valuable. The AI is able to discover useful information in term of connections. We like the easy method of creating a dashboard, and the ability to see the real user experience with the new functionality of the replay.
In the next release, I would like to see some new reports and more tiles on the dashboard. I would also like to see the Synthetic Mobile application improved.
Our primary uses for this solution are Production monitoring and RUM (Real User Monitoring).
Since implementing this solution we have gained stability, with 100% uptime.
The most valuable feature is the RUM. Finally, we have 100% visibility on website performance, including third-party sources.
This solution would be improved with the addition of annotations for automated custom metrics creation.
Our overall stability has improved with this solution.
The primary use for this solution is for monitoring and performance challenges.
Using this solution has increased log traceability.
The dashboard is very useful, as you can monitor different parameters on the same screen.
The thread traceability is something that needs improvement.
This solution is primarily used for performance analysis and problem-solving.
Dynatrace helps us to improve overall performance and allows us to detect the root cause of an outbreak on our systems.
The most valuable feature is the indication of root cause at the moment there is a failure in our system.
The configuration of this solution is quite complex.
Our primary use case is to helps solve Java bugs.
This solution has helped us to improve application performance and reduce issue-impact with faster resolutions.
The most valuable features are Root cause analysis and User session replay.
Two things that can be improved are the licensing and the Business dashboard.
We use Dynatrace OneAgent as a solution to monitor applications, as well as infrastructure. This idea was appealing to us for time management and efficiency purposes.
Since we moved to OneAgent, we are much quicker to address live incidents and problems that occur in our systems. This has given time back to our Ops teams to fix problems rather than spend time looking for the root cause. This had been problematic for us in the past.
So far, Problems grouped in one place is a personal favorite. This allows me to stay on top of reoccurring issues and drive teams to minimize those. This constantly improves our services. The Problem history also helps keep track of previous issues, should we be required to look them up.
So far, OneAgent has been very useful for us. However, there is room for improvement. More integrations could prove very beneficial to us. One example is more AWS Redis integration, which would help us monitor our digital estate more efficiently than we do today. Licensing and user management could also be improved.
We were previously using Dynatrace AppMon, which had limitations to monitoring applications. We were then demoed OneAgent, and it better suited our needs.
This is primarily used for Dynatrace Customer Managed Configurations in a closed customer environment. The operating systems include Microsoft Windows Server and various Linux distributions.
The Dynatrace solution has an important role in optimizing application code and SQL queries for customer solutions. It is also used for identifying, alerting, and reporting on real-time problems.
The most valuable features are the Smartscape, Resolution Path, synthetic and availability reports. These give a deep understanding of the environment being monitored, as well as addressing problems in terms of playback, code analysis, database calls analysis, and the system as a whole.
For an easy view of global and entity-specific configurations, a separate tile or pane aggregating these configurations should be implemented. This would make managing global or entity-specific configurations easier, as it would show previous configurations made to the environment by another DT user.
This solution is very stable.
This is a very scalable solution.
We did not use another solution prior to this one.
We did not evaluate other options before adopting this solution.
We are using this software to monitor an online banking system in a production environment.
Dynatrace has improved our organization because we can see big problems in the most important systems before these issues will be seen by clients.
The most valuable feature is the PurePath analysis. We can see each session, end-to-end, and discover issues.
The user interface needs to be improved.
APM monitoring and troubleshooting for any kind of technology.
This has improved our organization because any kind of technology, even legacy equipment, is now monitored.
The user experience monitor is a real added value.
It would be nice to have a simplified monitoring feature for non-Java applications.
We run our application on a Windows 2008 to 2016 environment. We needed a tool to give us insight into how our applications were running. Dynatrace was the answer.
Dynatrace gives excellent insight into the performance of our applications. This helps the IT department to respond quickly to problems because the root cause is reported by Dynatrace.
We find almost all of the features valuable, including Integration, Smartscape, and Problems.
The extending of Dynatrace with plugins can be better.
We are a partner and are selling this solution. We also use it internally.
This solution has helped us to reduce the time to detect the root cause incident and also to avoid missing incidents.
The most valuable feature is the beautiful UI. The correlation between events and impacts is helpful, as is the automatic detection of the infrastructure.
Improvements are needed in the navigation and time-frame selection when browsing problems, sometimes, the time-frame set change and I need to re-set it again, plus I find the navigation bar on the top left quite cumbersome, when i'm analyzing problem and I want to go back to the previous view, for me the nav bar should be the solution but often, when selecting the last element in the bar, I do not go back to my previous page but to the "main element page" (Aka, When doing deep analysis the analysis components are not shown on the nav bar and the only way is to press the browser "go back" button).
We use this solution for continuous monitoring in our customers' environments, auditing, and troubleshooting.
This is an easy to use solution and our customers like it.
The most valuable features are the UEM (User Experience Monitoring) and the DEM (Digital Experience Management). It is also easy to install.
I would like to see more default dashboards included, and maybe more possibilities in terms of customization.
The ROI for this solution is fast.
As a system integration company, we provide it as an APM solution to our customers.
Dynatrace solved various kinds of Ops / DevOps issues for our customers including bad code, bad configurations, operational issues, etc.
The features that we find most valuable are automatic root cause detection, topology discovery, and session replay. Basically, all the features provided are valuable.
I would like to see an App store for plugins and extensions. This would provide a single point for all Dynatrace users to download and install product extensions from.
Exceptional. Very quick replies (sometimes in the order of minutes), very professional.
Case by case.
AppD, New Relic
Have Dynatrace or one of its partner do a PoC. You will not regret it.
Our implementation is not a single case. We are the local partner for Dynatrace in Romania and have very different use cases that differ based on our customers.
Dynatrace is being used from monitoring the end user experience to infrastructure monitoring and real-time business and impact analysis.
The ease of use and full stack, 100% visibility inside the monitored applications has really helped us improve, showcase, and implement Dynatrace throughout our clients' systems.
Ease of problem detection, alerting, visual timeline, and user-sessions are some of the best features of Dynatrace.
There are lots of features that Dynatrace provides that really help the business grow. From user experience, where you can view exactly what your users are experiencing, to AI problem detection, and root cause analysis where the IT department is really keen to use Dynatrace.
The web interface, in some cases, is a little ambiguous to use. From passing through different screens to go and drill down, to not propagating the filter timeframes from different screens.
Apart from that, all of Dynatrace’s features are a real plus to have implemented.
We used this solution in a production environment to perform stress testing and evaluations to help find bottlenecks.
We monitored user behavior and identified the funnel most used by our customers. In other services, we wanted to detect third-party tools with the wrong service.
The most valuable function is the ability to replay a customer session. It is very useful to know all of the customer details (device category, Operating System used, browser version, location of the customer, and connectivity).
We also need to know which areas of our site have the worst performance.
It would be an improvement to know where we lost our conversion funnel. It’s very important to share the same information with other departments of the company and converse using the same dashboards.
We would like to see an AI tool that detects issues with our site in real time.
Nowadays, we are not Dynatrace customers. However, we want to keep up to date with product functionality in order to evaluate the necessity of changing our current solution.
We use this solution for performance monitoring and identifying resource management problems.
We hope to improve our customer experience by making our webpage, native applications, and B2B interfaces function properly. In our environment, we have several databases and several types of code. The databases include Oracle, MSSQL, and MongoDB.
It is critical to have the proper tool.
One of the features that sets this product apart from its competitors is that it generates a solution. We really like the number of reports it generates and the information that is provided to assist with analysis.
This solution also has the capability of analyzing our messaging software.
We would like to see more third-party tools for training.
We carefully looked at all of the options on the market and currently, we are in the process of a pilot installation of Dynatrace. The initial results are encouraging and we are waiting for more data to make the final decision.
At this time it seems that Dynatrace has almost everything we need.
We use this solution to assist with resolving incidents in a banking environment.
This solution helps the bank to detect problems faster.
The most valuable feature is the workflow, which helps you to easily have an overview of the infrastructure that you are analyzing, without having worked with it previously.
The AI is not that intelligent and there are different places where it could be even more automated. For example, when you are looking at the code tree and you must manually click to expand the tree until you find the process.
Our primary use case is to monitor the full stack of our digital platform, including the user experience. This provides us with business insights.
This solution has helped us with faster identification of the root cause, allowing faster resolution and increased uptime.
The most valuable feature is the AI, which makes root cause identification much faster, and the support much easier.
I would like to have the ability to share live data with selected third parties so that they can see how their product is performing for our company.
We use the Dynatrace platform to perform Application Performance Management for our Kubernetes clusters and some other legacy clusters. We have many Kubernetes clusters in our company and we are connecting each to an on-premise Dynatrace platform so that we can analyze the performance of each one separately.
Besides being an Application Performance management platform, Dynatrace has many important features that are easy to use. This gives us more time to focus on business. From installation to page load analysis, passing by monitoring hosts. Everything works like a charm.
The most valuable features of Dynatrace are the ease of installation and management, as well as its extensibility.
Dynatrace provides a wide range of platform supports that can be installed.
The service should be improved such that it is more useful to its clients. For instance, detect repeated performance-issue-related root causes, optional log streaming for hosts, processes, or a group of processes.
It needs a dashboard for cluster events in general, and for Kubernetes specifically.
For the Kubernetes part, we have a good operator that takes care of updating the agents so we have nothing to do.
Our primary use cases are for APM and digital experience management.
The ability to solve production issues quickly is the most valuable improvement for each of our customers. Also, rolling out and maintaining the solution is more than easy and requires very little effort from the customers. This is unique compared to other vendors.
The most valuable features are the speed of deployment without any changes to the monitored environment, and the very well defined and implemented user interface.
Working with the user interface is a pleasure for any customer since it is fast, responsive, and very well designed.
From the other features, the ability to drill down to the code level is often very valuable when doing an analysis of issues. Also, the integrated AI (assisted machine learning) helps to analyze anomalies without the requirement to define and describe your environment.
As the product is evolving quickly and product features are added on a monthly basis, a more transparent roadmap would be more than welcome.
Also, some parts of the business operations, such as pricing, should be improved since the terms are changing and it is not easy to do estimates.
We get APM requirements from various customers raising concerns about poor performance for their application. They do not know what is wrong and, most importantly, what is causing it. Dynatrace has helped a lot to identify performance bottlenecks and helped to fix them before the impact is increased.
Partnership with Dynatrace has given access to training, their knowledge base, etc. This has given us awareness about the latest trends in technology and how we can utilize Dynatrace to align with it. We have gained customers’ trust that we can provide them the best value from the Dynatrace solution.
The most valuable features for us include problem detection, root cause identification, Smartscape, and integration with cloud infrastructure like AWS, Azure, GCP, etc.
This solution supports integration with Gen 3 solutions.
The following features are also valuable:
The chat support definitely needs to be uplifted. Most of the time, chat support is not good enough for answering queries. Local support should also be increased.
There also needs to be more frequent training for partners, so teams can be enabled on product features and enhancements.
We use Dynatrace to monitor the API performance of the server.
Dynatrace has an auto-baseline and uses AI to monitor the performance of each API. The response time is related to the baseline.
The most valuable feature is the auto-baseline because I don't have to set it with Dynatrace. With New Relic, we had some problems with the configuration.
I would like to see the same features as in the New Relic Insights in the dashboard. That is the only thing I want to see improved in Dynatrace.
Dynatrace is quite stable. I don't have any issues with it. With Dynatrace we need to upgrade and restart often. New Relic is more stable than Dynatrace.
I have not had any issues with the scalability of either New Relic or Dynatrace. They are both quite scalable.
Currently, we get the vendor support directly. Their support is adequate.
Last time we were using another APM and the company discontinued the APM product, so we had to switch.
The initial setup of Dynatrace is quite easy.
We implemented the installation ourselves without any issues or problems at all.
Our most important criteria when selecting a vendor are the features and local support of the agency. Those two along with pricing are very important.
The reason we eventually chose Dynatrace was for the automatic (auto-baseline) of detection for the API performance issues.
We use New Relic as well in the same facilities.
I will rate the software a nine out of 10 because they are able to help solve our issues for us even though we don't understand the system fully.
To make it a perfect 10, Dynatrace needs to implement the features from New Relic in the dashboard so that I can monitor my own performance. Even though New Relic is not as good as Dynatrace, I have to understand my own system. I set each parameter manually before every launch by five minutes.
I would suggest to prospective buyers to evaluate both Dynatrace and New Relic to see which features are best for your company.
If you are not sure about the system requirements, choose Dynatrace. If you understand your own system and know by seeing a network outline exactly what you need for support, then choose New Relic.
Our primary use case for both Dynatrace and AppDynamics is for application performance monitoring (APM). The main reason for having application performance monitoring is, when we see something is running slowly, we can immediately look to see where the issue is at before our systems crash on us. So, one of the major roles it plays for us is the ability to keep our system performing in peak shape.
Our ability to see issues coming, then quickly isolate and correct problems was our main use of Dynatrace. We are not there yet with AppDynamics. It has been ten months, and we are still spinning our wheels trying to set it up and figure out how it works.
When collecting data with Dynatrace, we saw every single transaction that happened in real-time. Whereas, with AppDynamics, they take snapshots, and we only see a tenth of the information that we did with Dynatrace. While the information is there, if an issue with an application happened in-between snapshots, it would not be readily identifiable. You would have to go hunt and peck for it. We don't have time for that.
Using Dynatrace, we collected application metrics within three hours, in most cases. The majority of our triage were within three hours, then we were able to discover the root cause of issues.
With Dynatrace today, you have a single agent. You stick it on a server, and it doesn't matter if it's Linux, Windows, etc. It is a single agent and executable. You run it, and it injects itself into your collecting data. This is compared with AppDynamics, which on some of servers, we have had to install as many as four different agents and configure them all individually, trying to collect the same type of information.
The dashboarding for Dynatrace is ten times easier to set up and has more options of what you can put on it, especially if you are in a single payment class environment.
With Dynatrace in our environment, the managed server required root access to run. As a government agency with tight security, this has been an audit concern for us. A major area of improvement for Dynatrace would be to make it so the program does not need root access to perform. AppDynamics does not require root access to the servers. Once they are set up and configured, they can set their end run without root access.
The number one area of improvement for AppDynamics is to simplify their agent install. Instead of having four or five different agents to get all the different things that you need with different pieces of information, they need to figure out how to put theirs into a single agent, like Dynatrace has done.
We have not found AppDynamics in our environment useful at all. We are struggling to try and make it work. AppDynamics is for applications that are static. In our government agency, we are too dynamic. Everything is changing constantly, and AppDynamics does not work in this type of environment.
The stability for Dynatrace is probably 85 percent. We had enough issues where one of the services would just stop running, then we would have to restart it. Not very often, but it happened.
With AppDynamics we haven't been able to use the system long enough to determine its stability.
Dynatrace is very scalable. We grew it over the 12 years that we had it. So, that has proven that it is scalable.
AppDynamics is scalable as the environment grows. However, we are still in the product's infancy, so we haven't seen this happen yet.
A company with a single application over multiple locations, like a retailer, but only needs to worry about one application and monitoring it, this is the perfect fit for AppDynamics. If you have an organization with more than 40, you definitely want Dynatrace. AppDynamics is not a viable product for an organization with lots of applications.
On a scale of one to ten, my experience working with the AppDynamics onsite people and offsite support is maybe a five. I feel that they don't want to take responsibility for the areas that AppDynamics is lacking in. When things don't work in which they were sold, then they want to tell us that it is our environment more than their application not functioning correctly.
Whenever we had issues with Dynatrace, you could get Dynatrace support on the phone, and they were all over it. They would get into our machines, then take screenshots or look at the performance of the systems while it was running. They wrote custom patches to help us resolve issues that we had. I would rate the support that we receive from Dynatrace as a ten out of ten.
We have been using Dynatrace since 2007 and AppDynamics for just about a year. We have not been using them concurrently. Dynatrace was not renewed, and management decided that we would use AppDynamics. This decision was beyond my control.
With Dynatrace, the installation and setup were a piece of cake. It could be accomplished usually within fifteen minutes, and definitely, within a half hour of deciding to do it.
A big difference that we found between the two vendor is in setting the system up and getting them ready for production. With the latest version of Dynatrace, it took three days and we had it in production. We are still trying to get AppDynamics in production since last May.
With our Dynatrace system, it required three servers and the program when we installed it on the servers. It was straightforward. You just clicked, clicked, and clicked as it went through the setup, then you were done.
With AppDynamics, we are now on eight servers to make it function like it should. We are ten months into it now, and we're still trying to get it right. We have AppDynamics folks onsite to help us with it. It is just difficult to implement. There is so much to it.
I sat down with one of their architects to decide how many servers that we needed, how they need to be spec'd out, etc., and that is what we built. Then, when we stood it up, and it didn't work, some other personnel from the AppDynamics team came in to look at it. They said, "None of this is correct. You are on the wrong this and that."
Thus, we did not have consistent information and support from AppDynamics when implementing the system. However, it is up and running now.
Our annual costs were about the same for both AppDynamics and Dynatrace.
When learning Dynatrace, we brought in Dynatrace people to come onsite and take my team through a week long training. We did that two or three different times. They offered this type of training. They also have online training out on their community that I could set up for my team members. The effectiveness of that training was about 75 percent.
With AppDynamics, they have provided some online training. The take away from it (from my team) has been maybe 10 to 15 percent. The training is geared more towards sales than using the product for what it was intended. It boasts the features and selling points of the AppDynamics product instead of the ins and outs of how to use it once it has been installed in our environment.
I would definitely recommend Dynatrace. I have the benefit of having used it for so many years. It takes less infrastructure to set it up initially. It's a single agent engine. You just set the agent up and run it, then it configures itself. It goes out and finds all your processes with everything that's running, configuring itself. The simplicity of the infrastructure and simplicity of setting it up, then actually using it, along with setting up your dashboards to monitor your metrics is much better. There are more features than the AppDynamics dashboarding.
I would rate it Dynatrace as a ten out of ten.
At the point of where we're at with our AppDynamics experience, I would rate it as a five out of ten.
Dynatrace is an extremely helpful APM solution in large, complex environments, whenever we need to have an integrated vision of our users, infrastructure, and applications. It excels when we need to do all this with the lowest resources possible, with the best data quality, across all layers, and with the least management overhead.
The first time I came across Dynatrace, I simply remember the WOW factor. Being a performance enthusiast since the mid 80's, I cannot remember more than a handful of such technical experiences!
Dynatrace excels in troubleshooting scenarios. You do not get samples or gaps. You get the real thing. That is absolutely important when troubleshooting the strangest situations.
Focus on root cause analysis has dropped significantly for the importance of alerts. I was used to be drowning in alerts, but now have refocused. I still have them, of course, but now they are about minor things. I see myself getting rid of them very soon.
The most valuable feature is Dynatrace's commitment to bringing in new features. It is very difficult to keep pace with the new developments that are always occurring.
Given the full-stack approach, you are able to view from the end user perspective, almost down to the bare metal. It gives you all the way from the user perspective, network experience, web and application server view, application detail, process detail, operating system detail, and you even get to manage virtualization and datacenter support.
Dynatrace is capable of following all user actions across all monitored infrastructure, even to systems not monitored, but interconnected. This technology is known by the name PurePath, and being able to do this sometimes feels like magic. I have been told, time and time again, by very knowledgeable IT personnel, that this is not possible. Sometimes, you have to see it, to believe it.
One of the most promising features is the AI capability. I have seen some of its capability in production environments, and it is pretty impressive! It is not a marketing buzzword being abused, but it actually works. Being able to offload problem handling and root-cause analysis, and get to the point analysis of complex systems is something that I have not seen with my eyes, and in this depth, in any other IT solution to date.
Dynatrace's online presence in APM "education" is second to none. Viewing the YouTube online videos, hearing the PurePerformance Podcast, getting to learn in Dynatrace University, and following the documentation is like being in performance heaven! You just can't grasp it all!
In the new Dynatrace solution, support for legacy applications is still not there. Given there is excellent support of legacy applications and protocols in the Appmon & DC RUM offers, they have the knowledge to put it there. Knowledge that their competition simply doesn't have.
One to three years.
Not seen one stability issue since I been dealing with it.
No issues with scalability. The solution is multi-tenant and based in Big data technology.
What impresses me more is the ability to be able to analyze 100% of all activity going on, and do that with so low overhead. I have not been able to observe more than 1% overhead, despite Dynatrace saying that it can be slightly higher in some situations. I sometimes get the impression that overhead might be negative (underhead?), because you can get rid of inefficiencies like log dumping, that typically have high (disk and CPU) overhead.
Support is just great. Part of it is public, so everyone can check it out. Dynatrace direct support is even better, and you sometimes get direct answers from the teams that are implementing the functionality.
I have worked with other APM, IT, and UX monitoring solutions before. None of them is even near what Dynatrace has to offer.
Never seen anything so complex be so simple to install.
Pricing can be high, especially for Portuguese standards. But as one says, you get what you pay for.
Setup cost is very low considering that it is an almost totally, automatic process. Installing SaaS or Managed is only some minutes away. Given that there is no configuration involved in the agents, you can develop how many you want per hour. It only depends on your IT deployment strategy. TCO is thus much lower than expected. Licensing is very interesting, as you pay only for what is being monitored. A lot of things are given away for no additional cost. If you have a great IT consolidation, it will be pretty cheap. If you have a lot of servers, it will be heavier.
Since you cannot manage what you cannot measure, I do give the most importance to data quality. This is priceless. If you manage based on bad or incomplete data, you are leaping into the bad decisions direction.
The return might just be immediate. You install it and after minutes you are getting the full data in. The other day I compared Dynatrace to another APM solution, and the other person had been struggling nine months to get the data out. When he saw what Dynatrace did out-of-the-box, he simply could not believe it.
Finally, be prepared to be surprised! It is very fast pace. The Session Replay and the Augmented Reality are just two recent examples. Almost every day I get some new perspective in this field, like AIOps, and it just keeps getting faster!
I use this solution for application monitoring.
The functionality needs improvement.
We do not see any major impacts on the stability of the product.
I would need to explore this further because I have not tested it further than our needs.
We have not needed tech support.
It is on the high-end of the price range of products.
I advise that when you are looking for a new product, consider:
We use it to get the network stats and know how many clicks happened on both the front-end side and back-end side, then drill down on debiting to obtain the stats around that.
Dynatrace is working fine right now. It is working as we expected.
We can dig into Dynatrace to analyze data, then know where the user is going based on user navigation.
Performance-wise, we can see if there are any issues. Then, we can dig into them.
It gives complete stats of the user and what they are doing.
Because we are financial, there are certain things that we cannot put on the cloud. However, that is a given fact, not only for us. It is a given fact for any financial company because of PCI compliance. Because of PCI compliance, companies don't take the risk of putting data in the cloud. Otherwise, we have had a very good experience with the cloud.
The stability is good. I have not seen any issues.
I would rate scalability at least an eight or nine out of ten.
Everything is excellent with technical support. Some of their personnel are our main point of contacts. They are always in touch with us.
The integration and configuration is very easy.
Purchasing through the AWS Marketplace is excellent.
We did not evaluate any other products. Top management said. "Just go and use this."
It functions well. We are getting good support. It gives us everything that we were looking from it.
We use the on-premise version and have just begun onboarding the AWS version.
We use it for application performance management (APM).
It alerts us, or can detect, potential problems which are building up. Then, it let us quickly adapt our websites.
It is very stable and reliable.
We use a cloud version for everything that we look into, so we have had no issues. Scalability is working well.
The technical support is excellent.
We were previously using AppDynamics, then we switched to Dynatrace because it has more functionality, better additional components, and better management of problems. It also has a good AI.
The integration and configuration of this product were very easy.
It works quickly with all of our servers, databases, and load balancers. We are now testing it in AWS with AWS features.
It's helping us stay alive, afloat, and scale up as we need.
The pricing and licensing are very expensive.
Try it. It is a good product.
We have used both the AWS and on-premise versions. They are about the same for us.
The primary use case is for application performance management. So, we are using it to identify outages of different parts of the application as well as how we can make the application more efficient and rightsize it.
We can see down into the layers, such as with databases. We can see database queries which are causing problems.
We can see CPU usage for different containers. I can do a run and see what errors exist in containers which are causing problems. We can rightsize containers on the fly and understand what is happening with our Docker, microservices, etc.
The most valuable feature is it has AI in it. The artificial intelligence (AI) engine in it is able to do alerts and some good analytics. During outages, it is able to identify and correlate where the actual root cause of a problem is. This connectivity allows us to be able to see a bit further into the application down through the layers. If it is a problem within AWS, a problem within a container or something that a user did. We are able to see and coordinate that, then we are able to tell the developers how to fix it.
The GUI has the most room for improvement. Sometimes, it can be a little cumbersome to find things and be able to create your own views, or be able to dig in and understand where things are.
Some additional features would be the ability to break out some of the permissions and allow some additional or different ways to tag services, events, and different things which run. We want to push down the ability to do that, so developers and other folks can get in there. Currently, more permissions are needed to be able to do certain things, and we want more people to be able to use it, own it, and understand it.
We don't put very much stress on it. We could probably stress it some more, but we don't have enough systems right now on it to stress it. For the most part, the ships don't cause as much stress.
We are going to have it on about 40 ships around the world which will run it independently of our AWS platform. Those are don't stress it too much. We will probably stress it at a certain point, along with AWS, but we still very much growing the platform.
It can scale very well and very high. We don't need it to scale as much right now. It is able to absorb a lot of the systems that we have with the agents and and the API Gateways. It seems like it can scale very well when we need it to, so scalability is good for us right now.
If we needed technical support, we usually call our account team to help us figure out where the errors are, whether it is something with an agent or management servers.
It is pretty easy to integrate it into the AWS environment. You give it a username and password and it asks some basic permission. It can pull a lot of information very quickly. We are able to correlate more and provide more data for it. So, it was easy to integrate it into that environment.
We have it running on AWS. It integrates pretty well there. We have it on Red Hat Linux servers, as well as Windows servers. We have it running on VMware where it integrates very well. It understands these productions and understands our platform. It is able to read into Docker containers and all the databases that we run. However, it is limited as far as how many of a certain type of database that we can have, but for the most part, it runs pretty well and integrates very well.
It has been doing a good job of alerting us to issues. It has been very helpful and effective at identifying how we can do things to make our infrastructure and application a little better.
We considered AppDynamics, Datadog, and Crashlytics. We even considered things like Splunk for different pieces of it.
We chose Dynatrace because we needed something which could run both on AWS and VMware on our ships that might lose their Internet connectivity. This product gave us the flexibility of being able to do both. Dynatrace had the ability to run independently, so we could access it while it retains information.
A PoC is the best way to go. Put it against an application and go through the paces of tagging, analyzing, and alerting on it. You can understand what it does and how it does it. Give it a very complex application, so you can see how well it works.
We use the on-premise version because we have it running on VMware. We also use it on AWS to manage our systems on AWS for production and for our non-production environments.
Our primary use case for the product is going through logs and tracing through what has happened.
In general, it has helped me go through different logs more easily when something breaks.
Being able to quickly go through logs and figure out what is happening.
It needs a better way to figure out how to dig deeper into the details, e.g., sometimes we have to wade through multiple logs, etc.
It has been stable. It doesn't break.
Its capabilities are good.
We have a couple AWS accounts that we are running on the cloud. Then, we have a lot of on-premises applications, as well.
The technical support is readily available.
The integration and configuration of this product in our AWS environment is pretty seamless and easy to configure.
We don't really integrate it with anything else.
Purchasing through the AWS Marketplace was a pretty straightforward process. We had no hiccups.
I think the pricing is at a fair value for what it is.
We looked into New Relic and other logging solutions.
After doing our research, we figured out that Dynatrace was the best for us.
It is a very useful product.
Depending on your use case, try all the solutions out, then figure out which one is best.
We use Dynatrace for application monitoring.
We use this product with the goal to improve our abilities.
The most valuable feature is the monitoring of application performance.
I would like Dynatrace to be more flexible and user-friendly. They could do more in-depth analysis of the application monitoring. Also, they also could do some type of anomaly detection with positioning and have better integration with AWS.
So far, the scalability has been okay. Our customer base is hundreds of thousands.
The product needs a lot of the support, especially on the consulting side and post sales. You will also need an administrator.
I was not involved in the decision-making process. However, there are more competitive products on the market, which are more user-friendly, feature-rich, etc.
Research into similar products.
It adapts well for integration with other products in our environment.
We don't use it for AWS.
We are using it track how the data flows throughout the entire AWS environment and seeing if there are any road blocks in between and trying to fix those. Dynatrace is really good at illustrating those. It provides a nice graph, and we can see where everything goes. It is easy to explain both to the people that work with it and the customers who want to see where their data is going.
In terms of tracing where the data is going, some clients don't want it to hit a particular instance, or even worse hit a database that is not part of what its designed for. Therefore, it makes sure everything is where the customer expects it to be. It is a great tool for that.
The graphical interface is helpful, as it illustrates things well for anyone who wants to know about the information it provides.
If Dynatrace could take out the controller that would be great. It is one less thing to install right now. Though, I understand why they would need it.
The less stuff that you have on the instances which are running on the actual apps themselves are better for people that watch user products. So, if it could go agentless, that would be great, but I understand why Dynatrace would need it to capture the points. However, every time we spin up an EC2 instance, we have to slap an agent on it and that is more work.
I would like them to make those agents and controllers as small as possible. That would be great. Or, if they could remove them entirely, that would be great too.
I haven't had any major issues with stability. Sometimes, the controller feels a little overloaded if you have a lot agents running, but that's just a matter of sizing things up.
The scalability ties into the Dynatrace controller, because there is just one which talks back to your on-premise. While it is nice to have just one point to talk to, when you start having a lot of apps and things trying to connect to the same thing, it can cause some issues. I do get it, it is just a networking thing along with design.
I have not used the technical support.
It is a brand new environment. They didn't have anything before.
In terms of explaining to a customer how their data works, it has been a great tool. Instead of trying to draw it out, then hoping that is exactly where the data goes.
Try it out. They are other tools on the market, but with this one, the graphical interface is what I like the best. If that is what you really want, definitely go for it.
While working for an MSP, these customers' use cases vary - in most cases, we must do the fault domain to identify who is responsible in order to fix the root cause.
One use case: A customer blamed our data center, stating it caused broken sessions for their Citrix users. After looking at the network traffic data with NAM (ex. DCRUM), it was easy to identify issues within ISPs at Asia, and not in our data center here in Europe.
All of them which measures performance and availability. If you don't measure, you don't know. Looking at the data patterns and trends, you can also find out abnormalities in the whole ecosystem, not just in the target application.
Reporting and dashboards could be better, compared with competitors. However, Dynatrace has excellent API for data mining but requires extra reporting tool.
The primary use case is monitoring and diagnostics for production marketing usage.
The visibility into the application's performance helps the executives and managers by using easy dashboards. Our engineers are also super happy with the ability to drill down and fine-tune based on issues which we have seen. A bunch of issues with OpenID were easily investigated, then we were able to fix them quickly.
It has been a very enabling tool for us, especially for my team. The visibility that it provides through the application's behavior allows us to find trends based on our customized metrics.
The new Managed Edition is too complex. I feel like a fish out of water. From the on-premise version to the AWS version, our initial use has been very complex.
For the integration, I use a hollow testing tube called Performance Center. I would like the ability to integrate with it. This would be a good feature. While I believe it is there, it needs to be fine-tuned.
I'm pretty impressed with the stability.
With the AWS version, you can access the updates through the browsers, not worrying about the tick line.
With the on-premise version, you need to use the tick line for updates. There are times when architects, who do not use the product constantly, find their stage and production options out of sync, then they need to have two tick lines on the same system.
There was an issue on-premise. We were trying to troubleshoot a production issue. We had to run a bunch of queries for different time frames to see where the issues were and how recently they had been seen. This crashed the Dynatrace server.
We are now moving slowly moving in installments of the AWS version because our environments are not large enough right now. So, we haven't tested it yet.
I have only heard positive reviews.
We went with Dynatrace because of its ease of use and it is feature-rich. It helps you to drill down into bottlenecks and find issues. When you have highly integrated systems, it gives you an extra lens through your whole ecosystem.
The Dynatrace team helped us with the integration and configuration in our AWS environment.
There have been many advantages in terms of production and issue resolutions.
The product is pricey, but it is feature-rich, which is why we probably haven't looked away from it.
We are also using New Relic. Our product teams keep explore new options to see what is out there.
I prefer Dynatrace over New Relic because there are better features.
I would recommend Dynatrace Managed because it has more features, and go straight for the AWS version because it is simpler to manage. It can also be accessed through the browser.
We previously used the on-premise version, but have switched to the AWS version, which has more features.
Our primary use case is operations monitoring.
When something goes wrong, we have visibility into the system, can find the issue, and quickly get things back up. This was previously much harder to do.
It gives us visibility into the product and what we are doing operationally.
As we move into using more AWS native architectures, it should support everything that we want to do. We don't want to adopt another tool.
The solution has been really stable.
So far, scalability has been fine. We have not seen any issues related to it. It looks good.
We had a good experience working with their technical support.
We were previously using CA Wily (CA APM).
The technical support helped us spin it up, then we received training on how to use it.
We compared it to AppDynamics. While I did not chose Dynatrace, from a technical standpoint, AppDynamics and Dynatrace are pretty comparable. I liked how both of them worked. Because we were moving more onto the AWS platform, Dynatrace was more compelling because they were right there with us.
Kick the tires. Figure out how it fits your use case.
Our primary use case is application performance monitoring.
It helps with monitoring KPIs.
I would like them to add serverless capabilities, because everyone is going there.
The stability is rock solid. We put a lot of stress on it.
The scalability is amazing. It is the best. One of our customers is a massive healthcare customer.
I am a partner, so I know people in technical support who I can contact.
The integration and configuration was easy.
We have seen ROI with this product.
The price could be improved.
We also evaluated New Relic and AppDynamics. We chose Dynatrace for the hybrid version and price.
I want to tell people about its hybrid security capabilities. A lot of people have legacy experience with the tool, so it is valuable. They would not have to reinvent the wheel.
We use a hybrid environment, so we have to use the both AWS and on-premise versions.
The product is integrated with Splunk and ServiceNow. It integrates easily with them.
We log everything. Anything that goes wrong, we want to make sure that we are able to see the reason why. Therefore, we check metrics around CPU usage, RAM usage, etc.
It's a different way of thinking. Before, we were with a big, monolithic app in the beginning before fully transitioning over to AWS Services, which has been breaking it down into a microservices architecture. It has allowed us to look at it and debug out from a different perspective. Previously, we were going in and looking at server logs, logging into SSH toolbox and debugging manually there. It puts everything in one place, so everyone has one center has one center tool to look at.
For what we're using it for, and having seen the other side of things, where we were debugging and looking into logs manually, then trying to run an analysis on them, it was a painful process. Looking at it from that perspective, it makes things a lot more user-friendly if you are using a tool like this. That is just been my experience with it personally, which is nice.
It is tool for the job. It does what it's meant for and does what it is supposed to, which is good.
From a debugging perspective, when we look at things, we want to ensure that we know exactly what is happening at a certain point in time. It provides us a reference for being able to go back and look at data at a certain point, analyze it, then determine if something was the root cause.
When compared with other tools, the experience needs improvement. I would like them to build out the interactions and make them friendlier.
So far, so good. I know there are a lot of competing tools out there. For what it is and what we've used it for, it has been good.
I have never had any issues, so I haven't had to contact them.
There is time savings. People's times have been cut in half using this solution because we were previously doing a lot of that manual work. Now, it's a lot more automated, and the data is just there.
We used a different tools out there, like New Relic. Dynatrace is good, but there are features which other tools provide that it doesn't.
Do some research. There are a lot of tools out there with a lot of features, which people have bought into it. Make sure to get the right tool for the job. When you do bring a tool on, take it for a trial run first, then see if it is giving you the value which you are looking for.
Our primary use case is monitoring.
Previously, some of our web publications were so hard to find that we would be constantly monitoring it using a graph and a type of installation tool. Now, it is very different. Dynatrace has reduced the time it takes to detect the service failing in production.
They should include more mission learning into the product and provide additional performance metrics for application learning.
The stability is good.
Scalability is good.
It reduces our efforts to identify services failing in production.
Our experience purchasing through AWS Marketplace was good.
It is perfect for application monitoring.
The integration and configuration of this product on the AWS environment is good. We are using the on-premise and the AWS versions, which are pretty much the same.
I work with a product called Rancher, which integrates really well.
Primary use case is EMI, which is application monitoring. Our enterprise management infrastructure is supported by Dynatrace.
Dynatrace has solved our problems.
The view it provides for default analysis is very nice. The way in which it showcases how the metrics have been captured and how lucidly that they are displayed. This is a good thing to have from a technical and non-technical perspective.
For the manage services, they work on CloudWatch logs and are given CloudWatch logs only. I would like more collaboration with AWS and insight into CloudWatch services. This would be valuable, especially when detecting the fault of the root cause analysis. It would make the process go faster. Essentially, more integrated services with AWS would be of help.
It is stable. We have not lost data.
It has quite a sleek architecture with respect to the number of instances that can become an agent.
As for its environment, Dynatrace is ready to scale. Our environment is huge. We have EMR clusters ranging from 100 to 200.
We regularly connect with the technical support and obtain input from them. They are nice to work with, so we have been happy with the service.
We worked with architects for the best way to configure our Dynatrace in AWS. We selected the managed architecture, and there are less configurations and costs of adoption with Dynatrace.
While it is quite good in respect to its functionality, there are few area in regards to pricing that they can look at how to possibly change. I have heard it's costly.
We did evaluate other vendors, like Datadog, who were also good. However, Dynatrace was implemented earlier, and we continued to use it because it was satisfying all our requirements.
Our requirements include:
Dynatrace is pretty good as they are the market leaders.
We started with the on-premise version. Now, we are moving onto the AWS version. From the perspective of analyzing Dynatrace, it was able to do the EMI for all our data services. It has worked out well. We have been happy with it.
It can monitor your entire infrastructure on AWS. I don't see an option why you should not use this product. If you don't have AWS as a requirement, then maybe re-evaluate. Otherwise, I am confident in the product.
Dynatrace is a 21st century APM tool designed and developed keeping next generation technologies in mind. We implemented it in our dev environment first, and the results were awesome. We got firsthand RCA as soon as we finished implementation.
We are in implementation and adaptation phase. It would be very early to comment on this, but we are very hopeful.
The problem evaluation feature is an awesome idea, but bit difficult to pick up initially. Please make it a little more intuitive.
Dynatrace is nextgen tool.
We use Dynatrace for all our client-facing or student-facing core web systems, which give students access to their information. We are monitoring all those for user expectation and user experience management. We are also using application performance management metrics to detect and troubleshoot issues which pop up from time to time because it is a job. It's a Java-based environment that we're monitoring.
Dynatrace AppMon has served us very well. We are able to get insights into our systems, which previously took us weeks to be able to detect. It gives us a much better view on the performance of our environments.
One of the big benefits is being able to manage user expectation to better understand what the performance of our systems are versus what the users expect of those systems. We are also able to scale.
We can prevent major system downtime because the system uses baseline monitoring. It can when something is about to horribly wrong and affect systems in a short while. This has helped a lot. We are more pro-active instead of reactive. Reactive monitoring in IT is always viewed in a bad light. People don't like it when you just react to problems.
Keeping a record of full user transactions which we can then go back to the SIEM. Mostly, when a user complains about poor performance, it's very difficult to put a metric on it. With Dynatrace, we can actually go into the user's transaction and look at all the transactions the user has. We can see the actual metrics behind those transactions and what caused them to slow down and have poor performance.
If it's a case of the user's ISP and it is not actually our system that is the problem, it has given us a lot of capability to provide feedback to users, and say, "The system was slow because you were working from a slow ISP connection," or "You were working from a degraded browser." This helps specifically for cases where we want to educate our users on how to get the best performance out of their systems.
There is still a certain amount of technical skills needed to be able to understand what you are seeing on it. You also need a large amount of technical or infrastructure skills to understand how and where to install it.
The reason why we are looking at Dynatrace OneAgent is because Dynatrace OneAgent is better at troubleshooting than AppMon. Dynatrace OneAgent now comes with analytic engines and an AI system which helps you troubleshoot quickly. It also does root cause analysis. Therefore, we wouldn't need to do root cause analysis anymore, since it would show us the exact point where things go.
The capability development and user experience management that OneAgent would gives us is a step above what we have with AppMon. For example, we can see exactly what our full user compliment is leading us towards. With AppMon, we could determine:
Patching and updating is very easy. The system's stability is good.
Depending on how much storage you allocate, you can actually keep quite a lot of information in regards to your system. You can scale it yourself. The system doesn't force you into a specific storage medium.
It's able to scale linearly and vertically. It doesn't matter how many systems you have. You can just plugin more collectors and agents. It does this very well.
Every now and again, we contact technical because we have a few questions and they are very responsive and helpful. If it's a problem that we cannot figure out over the phone, they will make an appointment. They will come to our site. It's all part of the support contract, and there is no extra charges for it, which is good.
Previously, we were using infrastructure monitoring to a large extent. We didn't have application performance management. We realized that we have been spending a lot of time trying to figure out why applications were not performing correctly. That's the main reason why we went for the AppMon solution.
The initial setup was a bit technical.
We've used technical support mainly for the original setup.
The time that we save troubleshooting or finding the actual issues within the application execution.
The vendors on our shortlist were AppDynamics, CA, and Dynatrace. We chose Dynatrace because they performed the best during our PoC trials. It was the best all-round monitoring platform.
Make sure that you understand the scope before you start looking at application monitoring. Understand your environment.
Most important criteria when selecting a vendor:
I installed the demo application on an internal server and looked at the functionality of Dynatrace.
Through the demo server, we were able to see the functionality of Dynatrace and plan to apply it to our internal systems.
Through end-user monitoring, we were able to measure the user's perceived performance and build an SLA based on that information.
We hope the next version will have more powerful database monitoring capabilities.
Our primary use is application performance monitoring and real user experience. Our Dynatrace application monitoring has been in since 2012. It is performing extremely well. We have not had any downtime or issues with stability or scalability.
The benefits we receive using this tool increase productivity, which increase revenue for the state. A huge benefit of having Dynatrace AppMon in our environment is the proactive monitoring it provides. This help us avoid unexpected outages and downtime.
Its ability to deep dive into the application code and find bottlenecks that reduce productivity for the users and downtime. Proactive alerts are extremely beneficial and help us keep IT out IT team small.
The AppMon solution that we are using is the Dynatrace AppMon. I am currently working to upgrade it to the Dynatrace Managed solution. This is basically leaving AppMon and going to their next generation. This will streamline everything: Ease of installation, ease of use, and built its own intelligence, which I like to call self-healing.
Dynatrace AppMon is a tested and stable product in my environment. The only downtime I have is planned for patching servers.
Dynatrace is highly scalable and works well in our hybrid environment.
I would give Dynatrace's technical support a 100% rating. I feel like whenever I call or send an email that I get the right person automatically. For the most difficult answers, the most I have to wait is about three days and the answers have been relevant.
We have used Wiley from CA. It did not perform the way we wanted it to, which was a driving factor for switching over to Dynatrace products.
The initial setup started a year before I joined the team. I have been involved in the upgrade processes and they were straightforward.
Our implementation was with the help of Dynatrace. We wanted it to be fast and right the first time. Success on both counts!
I do not have dollar figures, but if I did, the ROI would be at least 100%.
Look at the product and the product features, not the price. Too often people look at the price and turn away. Dynatrace costs a little bit more than the other products I researched, but it can do far more. Since my last review, I have stood up a competitor's product. My Dynatrace installation is two servers plus my collectors. The competitor's product required seven servers. That is significant when looking at the cost.
I feel the price is good for what the product does.
I have done research on other products that are in the same market space. They cannot provide the same in-depth detail that Dynatrace does. I have since implemented, as a proof of concept, a major competitor of Dynatrace. The result - I will never stop using Dynatrace.
If I had just one solution that could provide real answers, not just data, the immediate benefit for my team would be less resources needed. This would streamline and automate things.
Most important criteria when selecting a vendor: reputation of the vendor. We go read reviews. We also check vendor references and talk to other customers to find out what their experiences have been.
The primary use case of this solution is to investigate performance bottleneck issues.
It helps our organization identify potential problems by doing thorough analysis of systems which integrate with one another.
It could improve its GUI interface. The GUI design is too crowded and the icons are small. Sometimes I end up clicking on the wrong button.
Product is very high tech and advanced on the one hand, and very easy to implement and maintain on the other.
We are not blind anymore with our digital services performance; no more fire fighting. IT Ops is now proactive and collaborating with our development teams.
Time to value was surprisingly fast.
Dynatrace AppMon has allowed a deep dive review of performance problems in near real-time for our primary external website and related web apps and web services.
PurePath view of methods and the call stack are extremely valuable for troubleshooting and performance review.
Dynatrace is a rapid release product, so new features or support for newer tech are being added all the time. Our primary wish list for RFEs or feature requests are additional integration options with ticketing systems. Although, we are able to work around it, 'ticketing' is not a core function of the product.
Releasing a new product and going from an AS/400 to a Microsoft environment.
We obtained a better insight into our environment and consolidated a lot of our old apps into one app.
We are still implementing Dynatrace. We are in the process of a PoC to discover why our sync test failed. It also gave us the cause of what was at fault.
PurePath gave our developers some tools that they did not know existed, and they gained a faster, more robust use case.
It needs education and training to ensure you get the full value of your purchase. Maybe add in a certification for Dynatrace.
.NET core support to the level of Java (at the moment, it is limited).
Custom reporting capabilities should be extended, because it now has basic charting capabilities. Alternatively, Dynatrace can create a bunch of plugins to popular BI platforms (e.g., Microsoft Power BI). All to allow custom reporting as well as SLA-oriented reporting.
Our primary use use is to monitor apps in terms of performance and availability.
We can analyse problems more quickly, and detecting problems becomes easier with Dynatrace.
PurePath: The transaction structure can be seen by any user.
Under heavyweight, Dynatrace becomes slower when listing PurePaths.
Searches should be faster.
It looks nice. The service discovery and user play are really surprising.
Use case monitoring of our shop floor systems.
We use it for acceptance and production environments, which are SAP based.
We can report and monitor specific use cases which could not be monitored with SAP or other tooling.
As of v.7.0, it is possible to use the web UI. Before that, it was too difficult using only a fat client
UEM can be used for user impact analysis and troubleshooting. We have applied this to prove that a specific issue originated from the back-end and was related to a specific user function.
Bring the interface to the same level as OneAgent. v.7.1 is a good improvement, but it has not been integrated into our environment yet.
One to three years.
We have all our infrastructure in the cloud.
This tool helps us gather information from all hosts and services, then cross reference the information.
Although, we are still implementing it, we have some servers using Dynatrace. We have been using it to find bugs in our production environment.
The cloud integration, because it allows us to monitor serverless services as well as Docker containers, etc.
I also like the "session replay".
They could improve their price ranges, as there is no option for startups or testing.
It has all the things an enterprise needs.
The primary use case is for application performance and analytics. We use it to monitor our on-prem Windows and Linux apps as well as our AWS and Azure cloud instances.
We used to rely on multiple operations tools to monitor and obtain bits and pieces. Now, we have it all in a single pane of glass.
It needs certain UI changes to make going back to certain Windows easier. Certain windows open up in a different category with different set values and throw you off if you are not used to it.
Our primary use case is performance improvements and understanding the customer struggle. Dynatrace gives us real-time data that we can use within the office to display data.
Provides more visibility into applications.
Customers want to monitor .NET applications and real user monitoring on online banking applications for desktop and mobile.
Dynatrace shows the customer path, common errors on desktop and mobile, and allows us to achieve faster page loads.
PurePath does deep dive analysis, has dashboards, and provides real user experience monitoring. It has allowed us to do analysis which was never possible before.
Our core application is still running on a Java Client-server based architecture. We provide that application as a service by hosting it (hybrid cloud approach).
We needed a tool that would provide insights into the application and support troubleshooting activities in order to reach our SLA objectives.
It has delivered on its promises and is now the single trusted source of information for various actors of the service support processes.
Its ability to correlate a large source of information to pinpoint a root cause. This speeds up issue resolution and allow us to better reach our objectives.
Dynatrace SaaS still lacks configuration API or command line which would allow moving configuration from one tenant to the other.
We mainly use the tool to do performance monitoring of our customer environments. Internally, we check how our software behaves when we are developing it.
Developers love the tool, because it is easy to use. They can immediately see how their code behaves.
PurePaths are the best. We can see everything that we need from them. Problem detection is also giving us valuable information.
Waiting for the session replay, because it seems to bring the end user interactions available for the developers.
We use it to see a clear view of the system overall and monitor applications effectively and proactively. We also use it on our core software.
It collects and analyses information with AI, which is useful.
Session recording is one of the innovative features, which could be very useful for developers and the marketing team.
Regarding features, it would be good if there would be some features regarding app security.
It is quite expensive for startups.
We are using Dynatrace to monitor our core application of the bank system. It is used especially in technical metrics monitoring.
Dynatrace improved the resolution of problems by making it easier to find root causes and easier to predict bad system behaviour.
The web dashboards are quite useful, good looking, and easy to use.
The NeoLoad plugin is awesome, and it gets results from load tests correlated with test scenarios.
C language integration requires manual implementation through the SDK, which is rather difficult and time consuming.
Performance monitoring of business critical applications in production and pre-production environments.
Mean time to recover (MTTR) has reduced significantly during major outages due to specific data pinpointed by DT applications.
Smartscape display for ease of visibility, pinpointing a exact problem, and providing necessary details for fixing and even improving on.
Log analytics in the classic synthetic and RUM tools would be a great addition. Although, it is understandable that this is offered as a good reason for migrating to Dynatrace.
Full insight on app bottlenecks, mainly in production. This is very useful for the development lifecycle, too.
Dynatrace is a powerful tool for full stack monitoring with APM capabilities and AI root cause problem analysis. This helps our customers with their digital transformation and gaining insight into their infrastructure.
Monitoring production environment at the hardware and application performance level
for evaluation of performance during testing.
Make it easier to define applications. E.g., provide an API for applications defined by REST services.
We are implementing this as part of an overall end user experience and application monitoring. Our environments include Java, .NET, web, mobile, Linux, Windows, etc.
We would like to see more dashboarding capabilities and the ability to export custom reports. Additionally, we would love to see more advanced log analytic capabilities, though the current ones are already ahead of competition.
Monitoring front-end for mobile banking within AppMon. We are using a balancer on NGNIX with a Dynatrace plugin to check availability by parsing logs.
Dynatrace helps to build business and non-business dashboards and the appropriate alarming on mail groups.
Native Java integration and building complex PurePaths between the front-end and back-end databases.
It provides a better understanding of what is going on. It helps developers fix old and new problems, helps businesses to understand conversions, statistics, and service health.
We are monitoring business critical applications which provide prepaid vouchers for different customers.
We can see all the degradation of services in real-time, then we know exactly what the root cause of degradation is.
Service discovery with artificial intelligence automatic anomaly root cause detection, and problems replay.
Web banking monitoring: With Dynatrace behind my customers' businesses, we are focusing on real user monitoring, dealing with user sessions, dashboards, and reports for their business.
Better root cause detection and improve root cause categories. In some cases, the root cause points out only a clue of what has happened.
Enables fast deployment of application and host monitoring of our estate with minimal user intervention and hosting costs for the tool.
Currently, we are still in the proof of concept stage, but it has massively addressed time to deployment issues. The integration into ServiceNow has provided key data for populating our CMDB to include service mapping.
The ease of deployment via OneAgent, which then finds what is installed on the server, is fantastic. In addition, Dynatrace provides probable root cause for any issues which assists our support team to quickly address issues identified by Dynatrace.
The plugin architecture is not very flexible, which makes it difficult to add the custom monitoring not available through Dynatrace, specifically around file monitoring.
The proof of concept has been so far successful.
Support from Dynatrace is excellent. They are always on hand for any queries, demos, and/or issues.
We used Dyntrace AppMon. Though it provided a lot of useful information, it was very difficult to deploy, which is why we are looking at Dynatrace.
The initial setup was very straightforward, as it is a SaaS solution.
Dynatrace SaaS was setup by Dynatrace, however deployment of OneAgent was in-house and very straightforward.
The setup costs for Dynatrace are low, however licensing costs are high.
We also evaluated AppDynamics.
We use it mainly to debug, control performance and KPIs of a tourism company. We can provide dashboards to business.
With Dynatrace, we can be more productive and agile. It allows us to be more accurate when we need to work with bugs.
The ability to go deep into production code. With this tool, we can work with bugs like no other tool can.
The dashboard tool needs to be improved. We need more options, because the look and feel is too old-fashioned.
Identifying performance bottlenecks on load testing environments. Before launching new code into production, it is imperative to know if performance issues will arise.
It prevents (together with performance testing) production performance issues which usually result in customer complaints or system downtime.
The product capabilities all together are valuable. You get a good insight into what is going on inside your code.
Application monitoring on our clients environments, including microservices, JVM, .NET, and many other technologies
Real user monitoring is one of the best things in solution. The possibility to analyze any particular user session is wonderful.
The product is really fast to implement and gives customers instant value. This is important, because it reduces costs on the implementation.
We are implementing software on our clients' environments
We use JS, JVM, stack, PHP, and Python. We use database and NoSQL solutions, such as Redis, MongoDB, and Hadoop.
Dynatrace is a good monitoring tool. It helps DevOps find all its problems easily and analyzes performance problems.
Dynatrace united some monitoring tools, such as app monitoring, JVM, PHP, network, log analyzer, etc.
A role-based view and a Python monitoring tool would make a simple user interface more usable.
Dynatrace provides visibility into the application and its performance from the user to back-end services.
Details of EUM: It helps the entire team, including our management, our application owners, and DB owners.
Adds value to application owners, DB owners, and provides visibility on how end users utilize browsers and where they are originate from.
The most valuable features are end user visibility, Smartscape, and the entire visibility of our data center, including SQL queries.
Cloud monitoring and reporting need improvement, as well as how to manipulate data and export it to share with business executives.
It reduces time and provides detailed info, showing problem correlation, and a single point of diagnosis.
With nothing more than three commands, or a simple dockable container, everything was executed in minutes. In one week, we had enough customizations to be ready for production.
It would have taken us at least two months to hire another person from SysOps to achieve registration, supervision, alerts, and APM implemented with cheaper or free open source solutions. It was much cheaper and faster to go with Dynatrace.
We had users in a remote office complaining about the latency of the application and were able to determine the problem derived from the configuration of a router with the help of Dynatrace.
Dynatrace has designed its agents to capture limited stack traces for each transaction executed. It is a product that helps developers, testers, and operations to make sure their applications work quickly and reliably.
We also love the automatic alerts. Dynatrace alerts are based on deviations from the reference metrics which are constantly collected.
The one thing that I do not like about Dynatrace is their Web UI dashboards are very slow. They seriously have to improve their Web UI dashboard configuration and SSL timeouts.
No other provider provides record ingestion, Kubernetes/docker monitoring, and application monitoring for Node.js. Some competitors offer aspects of these, and some offer all these, but not with Node.js.
Dynatrace was the perfect fit.
Dynatrace was an incredible find!
My primary use case is to monitor business applications, mostly with Web front-ends, to provide access for end-users and consumers.
We have reduced our troubleshooting times and improved the way that we deal with most of the bugs in the applications. It's very easy to reach the root cause of the problems in the applications, due do the analysis with Dynatrace. The timeframe to update and fix the applications has been reduced a lot compared to what we had before Dynatrace.
It's the ease of deployment and ease in configuration. It's very comprehensive in its features to monitor end-to-end transactions.
I'd like to see more agents ready to be deployed. I know that it's possible to develop integration with Dynatrace API, but in day-to-day operations it's hard to do that kind of customization. So if they had more agents for more platforms and more applications, I think it would be better.
We have absolutely no problem with the solution. It's very stable.
In terms of scalability, my company is not that large, but it looks like the scalability is very good. We don't have any problems with scalability.
It's very good technical support. We don't have many issues with the product but when we have, we get very quick solutions. It's very good support. Nothing to complain about.
We used multiple solutions, the ones that came with the different applications. So a solution that monitored the database, and another solution for the application server, and a different one for the server hardware, and the connectivity. We had to integrate all the information that came from these separate solutions to come up with a conclusion about what was happening. This was the main driver that lead us to look for a real end-to-end solution.
The business perspective had a lot of weight in our decision because it was hard for us to really correlate the components of the application, and how it impacted the business application and the business itself. These were the main drivers that lead us to buy this product.
In terms of the most important criteria when selecting a vendor, we usually look for a product or a vendor that has good positioning in industry reviews. With that, the pricing is very important. Also, the features. So it must be a top vendor with the best possible pricing and the features that fit our needs.
I was not really involved in the initial setup. I just coordinated the deployment. But it didn't need seem we needed to do too much for the setup of Dynatrace. We had to focus our efforts on group duplications, from a business perspective, but everything else was discovered and automatically set by the Dynatrace application itself. I don't think we had much trouble.
The real complexity that I've seen with Dynatrace is to learn how to navigate through all the options in the troubleshooting process. We have a lot of ways to evaluate the same problem. We had some difficulties in the beginning with the use of the product, but after some time and some experience we have overcome this problem.
I understand that due to comparisons we did, that Dynatrace is still kind of an expensive solution compared to others. But I recognize that they are ahead of the competition when we do a feature by feature comparison. We have a very stringent budget for an infrastructure solution. Maybe if they provided modules, a simple module with fewer features and a lower price, that would be very good.
CA, Computer Associates, was on our short list as well as BMC PATROL. These were the main vendors when it came for my evaluation. Some research was done and I received these vendors as the best options to evaluate.
We decided on Dynatrace, over Computer Associates and BMC, mostly because of the difficulties that we thought we would have after the setup of the product. Dynatrace was the most expensive, but we had almost no need for service, for Professional Services. We just did some training and we contracted some consulting hours and that was it. The deployment with Dynatrace seemed to be easier than the others. We needed to get results very fast, due to the size of the investment.
I would rate it with a grade of eight out of 10 because it's a very good solution. We got results very fast after the initial deployment, but I still find it very expensive. So we are still being questioned about the cost-benefit. The value that we have in Dynatrace - I don't think I will have this kind of budget in the near future - it's worth it now.
My main advice is to evaluate the effort to set up the solution, customize the solution, after acquiring it. I know that we had better pricing, lower pricing, with the other competitors, but as I talked to some other customers that use BMC and Computer Associates, everyone told me it was a long run until they reached the setup that they needed. And they still have a lot of maintenance. Every change in the thresholds of the applications, they have to come back to the standards and redo the setup, but Dynatrace does it all by itself.
Application process monitoring, and user-experience monitoring.
It's used for monitoring inside of applications, like JVM thread-level applications, plus the user. I monitor usage statistics, performance statistics, plus the user satisfaction level.
We got to know which modules of my application are used more by the customers, as well as which modules of my application are very slow. It has helped with my overall my performance statistics.
There are different phases of product development. In the testing phase I was not able to determine if things were going fine or what was going wrong in my application. But using this tool, I got to know - even before the customer got to tell me - this or that particular module was not working.
Using this tool, I got to know the moment he got a page-out error on his screen. It told me, for example, a person in the US is facing this particular issue. Because the alert came, we worked on it, we resolved it and things were easy.
The moment he called us we just said, "Yeah, we have already acknowledged that issue and we have resolved it. You can just try it again." The person was happy. The customer's satisfaction has improved, overall.
We also got know many internal code bugs which could have caused memory leaks or other issues which we were not able to catch during that development phase.
We also got to know, on the network level, where the latencies were. If you go via Google, what Google says is that you have a number of things on which you should measure your performance. One is if there's an error or not. Dynatrace tells you whether is an error or not. Second is saturation, whether something is getting saturated. You should be aware of what is getting saturated. Dynatrace even tells you that. The third is if there is a latency. Network latency is also told to me by Dynatrace. So these are three things I got out of Dynatrace.
The performance features are most important ones. That would include reporting, you get reports on your performance data.
The next is the alerting feature. For example, my applications are going out on some thresholds. So I get alerts, according to the thresholds I set. I get that data via emails as notifications. That is another feature that we primarily use and we like from Dynatrace.
Also, it's easy to use. The usability is better in Dynatrace.
The fourth thing is: my customer is facing some issue. The linking is very good in Dynatrace. What happens in other monitoring tools is the linking is not proper. In those solutions, a person has to manually link many of the layers and what is happening in them, while in Dynatrace you get that from the very first visit. For example, if a person is visiting your website, from there it will traverse you to the end. If the application is a Java application, it will traverse you there, to the Method level. So that linking and traversing is better in Dynatrace.
If Dynatrace is capturing everything in your application, it has to "sense" that information, and that sensing needs sensors which we have to include in our applications. The more you apply sensors - the more details you want - the more you have to increase the level of sensing. If I increase the level of sensing, my application's performance goes down, because something is there that is, again and again, checking each and every thing in the application. So that load on the applications increases.
So, many times my applications used to crash because Dynatrace was working on them. So that was a negative point. We had to remove some sensing; either we had to reduce the sensing or we had to remove Dynatrace immediately. So that is one thing we don't like about Dynatrace.
And the second is dashboarding. The dashboarding in Dynatrace is not very good. We have used other monitoring tools like AppDynamics. We are also using AppDynamics for some of our products. If I compare Dynatrace with those monitoring tools, the dashboarding is not as good. If I have to create a dashboard it takes me time, the experience is not that good. The automatically generated reports are good, but the dashboarding was something we were expecting but did not get.
Also, sometimes it happens that we are not able to capture things. For example, if a person is logged in from India, from the city of Mumbai, and is using a Chrome browser, and his email ID is firstname.lastname@example.org. But what happens is, Dynatrace just fetches two pieces of the information, not all of it. Sometimes it gets it all, sometimes it doesn't. So that also came into picture.
The last and the most important, which we did not like about Dynatrace - and that's why we switched to other monitoring tools recently - was the support. We were not able to get proper, good support from Dynatrace. We had to raise a lot of tickets and then, one fine day their people would eventually come around and resolve the issues. The input we had to give them was very high. Support was very bad for Dynatrace, especially in the India region.
When you go for a monitoring tool, there are two types. One is SaaS. The SaaS version means that the monitoring tool is not deployed on your servers, they are deployed on their servers. The second one is on-premise.
We were using the on-premise and we were using very good servers for the Dynatrace deployment. Stability didn't come into picture. The version that they gave us was very stable. There were was no code bugs. Actually, there were some, but those were fixed immediately.
As far as on-premise was concerned it was fine.
Scalability was fine. We were using around 100 application licenses, 100 Java licenses. On that scale it was able to handle everything. There was no reason to scale it up. I'm not sure if it could be scaled up to 500 or 600 licenses. We didn't do that. It was able to handle the load in the one tier.
I rate tech support very low. The technical support was not that good. They were not very attentive whenever there were issues, even the critical ones. We were not able to find the proper support from their end so I don't rate it very well. A one or a two out of five.
Initially, we were using New Relic. But, after that we switched to Dynatrace because of the amount of functionality, the amount of troubleshooting it was giving us was more. That's why we shifted from New Relic to Dynatrace.
But once we saw the negative points of Dynatrace, we recently shifted from Dynatrace to AppDynamics. We are in a process of shifting all applications from Dynatrace to AppDynamics.
It was straightforward. It was was easy to add up. It was not complex at all, while the other monitoring tools are more complex than Dynatrace. Dynatrace was not that bad.
The licensing for Dynatrace is high. If you want to go for monitoring solutions, then why Dynatrace? If you have a particular budget, you can go for many other monitoring tools - apart from Dynatrace - and they can help you more and give more data than Dynatrace can.
And secondly Dynatrace also comes with a lot of issues, which I have mentioned elsewhere in this review, which can easily be rectified using other tools. It's not worth the money that you spend for Dynatrace.
Initially we were using New Relic itself. Dynatrace was one thing that we were evaluating. ManagingEngine was a new one at that time from what I recall but, I'm not sure which ones were part of the evaluation, because I was not totally a part of the PoC.
Apart from that, if you want to use on a system level, you can use Nagios, that's freeware. It's also good but, again, it is just a system monitoring tool. It's not an APM. So if you wanted to go for APM, then only New Relic. It was the one competitor for Dynatrace, at that time.
In terms of implementation, it's quite easy, now that there are many automation tools. So just integrate Dynatrace with the automation configuration tools. Just ask Dynatrace which integrations it has, for example Chef, or Puppet. If you integrate, the configuration will be easy.
Also, the configuration needs to be standard. The standards should be set initially. There should be a standard protocol; that needs to there. If that's not there, then issues may arise later on. These are some things which are advisable when you work with Dynatrace.
I would rate Dynatrace a six out of 10. When I consider all the negatives plus the positive points which I have already discussed, I end up at six, including the licensing and everything.
Application monitoring to quickly troubleshoot production issues and determine the root cause, as well as non-functional performance testing in QA.
Dynatrace has helped us reduce outage times and severity of impact.
So far, we have not achieved the benefit of preventing issues.
Quick availability of multiple aspects of performance from infrastructure to application layers.
Experience with relationship/account manager has been really poor, it does not seem to be the firm's priority to support their customers.
Our primary use case right now is that we are dealing with the perfect days and non-perfect days.
We have a certain goal. E.g., for this financial year, we want to have 350 days of perfect days. The perfect days is where every application in the business should have low business impact, all applications should be available, and there are some other metrics that we want to know.
We have constraints whether this is a perfect indicator, and saying whether it is a perfect day or not. There are some situations where one of the JVM is down, but the other processes, which do the same thing are up. In this case, pretty much there is no business impact. However, there is a technical issue, though no business impact. We can't sell this as a non-perfect day. Yet, it is a perfect day.
Right now, we are using AppMon. Therefore, we are using AppMon to find out what metrics are available for us to see or indicate what is a perfect day or non-perfect day. That is one of the things. We are also using reactive stuff more for reactive stuff at the moment.
We are using AppDesk on user experience in indexes for customer satisfaction.
We have been using for almost one year. It has been performing well.
The most valuable feature in AppMon is the PurePaths. Previously, we did not have this feature. The transactions using PurePaths are a good thing.
We have like different teams like in a cross structure (DevOps, infrastructure, IT, product monitors, and developers), who recently joined the team. They are not aware of what are the missions their touching and what are the other components or services that they are depending upon. Because of all this, it has been very helpful for the developers who just joined the team. It can be explained, "Okay, this is our application. These are all the components that we are touching. Changing any of this might affect all these structural things. If we want to do any changes to this, we might need to put all this in our test cases, then QA it just to make sure we are not corrupting it."
In AppMon, the performance could be improved. That is the one thing I am most interested in.
The other thing is the database. They might improve the database stuff a little bit more. The metrics and whatever that they are providing for database.
AppMon is lacking the AI that can be found in Dynatrace Managed.
Maybe last year, we had issues interacting with the MQs and the mainframe. They completely resolved this issue in 6.5, so we are now good.
We do not have downtime using Dynatrace.
It does scale well.
Right now, we have some performance issues with Dynatrace AppMon, but the AppMon team is working closely with us registering the number of issues we are having and providing an extra set of tools that helps us to make the performance of the tool better.
One of the performance issue is when we are trying to bring up user data, such trying to bring up 20,000 or 50,000 PurePaths. That is where it is taking like five to seven minutes. When we are on a call, we do not have that much time. We want to make it approximately less than two minutes. They are doing a great job on that.
I would rate the technical support very well. They work with the inside their development teams to get us the best answer, as much as possible.
We have LoadRunner and Wiley. Our company has used Wiley for 13 years.
They are one stop tools. If there is an issue, we have a team that always is on call: One should come from infrastructure, another from the data side, another from the product AVR, another from mainframe, and one person from Wiley saying, "These are the threats we have and open systems." So, we have five different people on a single call for a single issue. Sometimes we have 10 to 15 people on the call to figure out the issue.
With AppMon, looking at the translation flows and the PurePaths, we can with one or two members can identify and start to find where the problem is. So, this is a good feature.
The initial setup was straightforward.
We did have technical support help. During the first year, they have given us a resource, who has helped us in setting up the Dynatrace: getting some applications onboard, how to set up on dashboard, etc. This has helped us a little bit getting more familiar with the app and getting up to speed.
We have some demos on the Wiley, but they came very late to the table. Also, Splunk ITSI reached out to us.
For 13 years, we have been using Wiley. We definitely liked seeing the PurePaths being more helpful for us. As to Splunk ITSI, there is more configuration than AppMon. That is the reason we chose AppMon.
If you are implementing it for mainframe or MQ stuff, what are the things available and what are the other configurations that you need to set up.
Right now, we do not have AI capability. We are on AppMon. For us, it is about going and debugging the PurePath and looking into what is the issue: finding out the other use cases or root causes. It is pretty much manual. We are trying to moving from AppMon to Dynatrace Managed within the next six months. We are planning to do a debug on that. Going through all the videos and classes, it seems like Managed makes more sense for us and would be more helpful than AppMon.
If I had just one solution which could provide real answers, not just data, the immediate benefit for my team would be escape being pulled into a call and spending most of the time in analysis finding the root cause. If we are able to find the root cause and fix it immediately, that downtime would be less. That is the biggest benefit.
Most important criteria when selecting a vendor: customer support. That is most important, because companies do not have the tool knowledge initially, and someone needs support it or they need to hire someone. For companies like us, initially we onboard someone who has much more experience with the application inside the company, because we need some training on the customer support: when to support and what we need to do.
The next one is writing the PoC, we have to find out whether it is satisfying all our use cases. So, if a system helps us with our issues, that would be great.
We’re using Dynatrace AppMon currently in our environment, and we’re using it for troubleshooting performance issues. We are mainly using it in the performance-testing environment, to try to reproduce a problem if we happen to see one in production, and find what the root cause is.
It's performing well, absolutely.
I’m not sure about the entire organization, but for my quality assurance, the mean time to resolution or to find a problem has been reduced dramatically now. I don’t have a percentage but, we used to take a week or more to troubleshoot an issue, now it can be done very quickly, probably in a day’s time.
PureStack, I just love it. It can give visibility from the end-user perspective right through to the code level. That's the most valuable feature.
Also the UI is amazing. We really like it.
I think Dynatrace is top-notch, it's well ahead of its competitors. I don’t see any features which another vendor or other products have which Dynatrace doesn't. I think Dynatrace is in pretty good shape right now. I don’t really have any features which I’m lacking right now, so it's all good.
In terms of new features, I’m excited about AWS monitoring, that Lambda function, and log analysis. We’re not yet on the cloud, but still it's a good feature. We are actually planning to move to the cloud, and my organization is actively looking for tools which can support monitoring. This will definitely be a value-added feature.
It's not a big concern for us at this moment.
We haven't used technical support.
For deep diagnostics we were using an HPE product or, then, a Microsoft product called Diagnostics. It was difficult to use that tool and connect the dots. It was per machine base, per JVM base, and was not really giving a holistic picture. But Dynatrace is doing that all for us. And PurePath, again, I just love that. That was missing.
I think one of the vendors helped us. We didn’t have any hands-on, but I think we did some Dynatrace University, we’ve been through some videos. And the vendor gave us some training, so we’re fine with that.
I’m a big advocate of AI. It seems that AI can join the dots sometimes for us, and that is helpful. Instead of spending the time to think and connect all the dots, AI can do that for us in the future, and can come up with a solution also. That will be nice.
In my previous company I did use siloed monitoring tools. We used HPE BSM, Business Service Management. There were two piece to that. One was infrastructure monitoring, host monitoring such as CPU, memory, using SiteScope. The other was end-user monitoring, synthetic user monitoring. Also, there was a piece called Diagnostics. The challenge was, although the two pieces, synthetic monitoring and the host monitoring, both were agentless, it didn’t give us the real root cause of issues. Diagnostics did but you had to go and install it, and it didn’t perform very well in production.
In terms of one tool, right now I’m seeing Dynatrace can do a lot of things: the entire DevOps, infrastructure, application performance monitoring, all that can be done using just one tool. Now they just released a new feature for log monitoring. I’m really excited to learn about that. If everything can be packaged in one, that would be nice. You wouldn't have to worry about different vendors and patches. And especially, they have a SaaS model. I’m a big advocate of SaaS. My company is not right there but eventually, I hope, when it gets there, I think you’ll see big use of it, and ease of use.
I think from the organization's perspective, probably the most important criteria when selecting a vendor would be the cost, and the tool, obviously. The tool is very important: quality of the tool, reliability, scalability, all those factors weigh in.
My advice would be, from the tool perspective, to look at this tool and its features: ease of use, scalability, and stability-wise this tool stands out. I understand organizations have a pricing factor, a cost factor. That is something you have to decide on. If you want a low-cost tool, there are different tools in the market, or do you want to settle with the best tool in the market but you'll spend a lot more money. Do your research, work with your peers, your leadership, understand which way they want to go. But definitely, as an engineer, I will always say you should go Dynatrace.
Our requirement is to monitor and manage application performance. Our primary solution is AppMon for application monitoring.
The performance comes with a little bit of overhead, but we are designing the solution that way. If my application allows to be within that overhead, we are using it.
I think one of the interesting parts is they are going to deliver, in the new release, artificial intelligence, which is like analytics that will improve the correlation technology: What is causing the problem? So that will help us to drill down the actual problem, more quickly, rather than analyzing from another tool's perspective.
For us is it is the PurePath technology which is helping us to drill down to what the root cause of the problem is. That helps us to escape the war room.
I would like to see machine learning, which will give it even more of an advantage. And the self-healing, which is there, I would like to see it even smarter, to get it to quick healing, itself.
Stability is very good. Initially, when some of the problems happen, say application issues come up, if you haven't properly designed, then it's really breaking. But now it is quite mature to handle those things, so it is quite stable for us.
Because we are now we are moving out of AppMon to the Dynatrace tool, probably in the future the scalability will be really helping us.
We use the Dynatrace University, plus some of the guidance solutions. Generally these guys are quite experienced and they help us to understand properly. They cannot help our environment - we help them to understand our infrastructure. It's like handshaking with each other. It's good.
We used a lot of different solutions. Those other solutions didn't have the ability to focus on a problem. What we had was application performance monitoring and not measuring as well. They were mostly infrastructure monitoring, but we really needed some sort of DevOps tool, which would help us to really know where our problem is.
I was the architect, I designed from the very start. In terms of the setup, initially there was a little bit of a learning curve, but now, because the tool has been simplified, I think this learning curve is gone.
When you adapt a new tool, obviously the people and culture need to adopt it, so it will definitely be a challenge. We found it was quite challenging in terms of the learning curve. But now, after two or three years, it is quite mature.
I think we had quite a lot of different vendors, but the reason we went with Dynatrace was because they were number one in industry reviews, and there were a lot of features we really needed that were already in place.
The role of AI, when it comes to IT's ability to scale in the cloud and manage performance comes in when the complexity increases, because your stacks are high, so human intervention would be really difficult. That's the reason we need AI power. Dynatrace has that capability and we want to use it. So looking to the future, AI and analysis will help us.
We have used siloed monitoring tools in the past. The biggest challenge is that we don't actually have one view. Let's say a new application is launched or different tools, they have a particular focus for a particular problem solution. But when you don't have a major overview, one view, it is very difficult to find the solution. So Dynatrace helps in that we can easily get one view, correlate, and everything will be one single pane.
If we had just one solution that could provide real answers, and not just data, the way DevOps is going on, we would probably want to adopt that solution. It would really help us regarding the cultural change of DevOps; the tool would help bring us to that level.
Our most important criteria when selecting a vendor are, first, do they meet our requirements. Second, whether they have the right support for us, when we have a problem can they immediately eliminate it for us. They need to understand us and to have the skill set to remediate those things.
I would rate it a nine out of 10. There have been some improvement announcements, but overall they are doing a great job.
I would recommend Dynatrace to colleagues.
The primary use case is application monitoring. We are using APM to test for performance, bugs, and hoping to resolve the issues faster, and hoping to catch them before we go to production.
It's been great, it has helped us a lot. It can do more, but it's definitely helped us a lot and I'm a big believer in Dynatrace products.
First of all it's mitigating issues before they are in production, and if they do go into production, it's reducing the time to find the issue and actually resolve it. I believe, in the organization that we're in right now, that is challenged for resources and time, a product like Dynatrace helps immensely.
The PurePaths, because that's where somebody who is a non-developer can figure out where the problem is and send appropriate PurePaths, clean charts, or even the link to the developer. The developer can then look at it and figure out exactly where the problem is, this is the piece of code that took the longest time, and then resolve it.
Right now, since I'm primarily an AppMon user, so maybe the Dynatrace product addresses this: The challenge with AppMon is, what if you don't have an AppMon agent on a host, but it talks to the database. It talks to it, but I don't have either a host agent or an AppMon agent on it. That has been a challenge, but I believe the Dynatrace agent, the OneAgent, will solve that, potentially. You ask me three months from now, after we take a crack at the Dynatrace product, maybe my answer will be different, but I'm hoping that addresses some of the issues.
The configuration of the alerts, that's been a challenge in AppMon for me, right now. Some of the alerts are too noisy, but that might be my lack of some configuration. Again, it's just me primarily handling it, so that could be an issue. Somebody asked a question in one of the sessions, here at the Perform 2018 conference, about noise and how many alerts to your problem count, and the person doing the session answered right away saying, "I checked my dashboard before I came to this session and I had one alert on it." So I'm guessing that will resolve itself.
I think I like the direction it's going in, the only challenging part for me is to keep up with the name changes. But other than that, as far as stability, I think I'm happy with it.
I think for the deployments we have right now, it has not been a challenge. My goal is to increase the usage throughout the organization, maybe that's where I'll face some challenges, but at this point there are no challenges.
I have used technical support in the past and they are pretty quick to respond. The other thing is, the APM community is available, Andy answers pretty much any question I post pretty quickly, so I think that group community help is really good.
I have use siloed monitoring tools in the past. When I started at Nemours 17 years ago, I had custom scripts that I would use to apply to various servers. They were on the host level, but the deployment was challenging. How to tie in a CPU alert to application slowness is challenging, because you had to go to the timestamp, look at the log and say, "Okay, this might be the issue." Dynatrace tells you how it is, and I think that's the most important feature.
I'm the primary Dynatrace admin, if you want to call me that, and it was pretty easy. But keep in mind, my skill sets are probably unique in the sense that I understand applications well, so I know how to insert the agent - because we use AppMon - how we insert agent into JVM.
But I believe the new Dynatrace product is probably the way to go, because you don't need to actually talk to the application folks, you just deploy it on the host and you're done. I believe it's definitely going in the right direction. It is complex, it wasn't for me, but I can imagine it being complex for some people.
I don't know if there were any other vendors on my list because we've been users for about 15 years. We started out with a Vantage product that moved to server monitoring, and then we had the Gomez platform, and then you also had Dynatrace, but then they all came under the same umbrella. So we never really evaluated any other vendor. We had some of the free tools we used to use, like Profiler, but from what I've heard from developers, nothing ever came close to this so I'm a fan.
If we had just one solution that could provide real answers, as opposed to just data, we could spend less time on troubleshooting and trying to figure out what the problem is, and actually do our jobs, which is to design, build, and develop software.
The criteria we look for when adopting an APM solution are ease of use, and does it truly get you down to the problem area - and I believe that Dynatrace does - and the third one, it's true for everyone, is the cost.
I would rate it a nine out of 10. I'm not giving it a 10 yet because I would like to see the Dynatrace product in action and truly want to understand it. If we move to Dynatrace, away from AppMon, are we missing out on something?
My advice would be go with Dynatrace.
I mostly just onboard different applications in the company under the Dynatrace platform. Occasionally, they will have issues, then I use AppMon in order to tell them what the issue is. It usually is something simple: The URL, this particular service is slow, or your database is not responding correctly.
It is performing well.
It is different for me than other users. I like the PurePaths dashlet the most. This is mostly because (and I can count a handful at times where this has not been this scenario) as soon I open the PurePaths dashlet and sort by response time, there is the problem. Every time.
If the PurePaths dashlet pulls up 750,000 PurePaths, I really only needed to know about seven or eight of them. Then, being able to look into the code-level dive about it, that is just a sanity check. Just to make sure that it is the same issue multiple times, not a random anomaly where everything else was crap.
Also Errors dashlet, I use that a ton.
I would love to see a better data export, because AppMon's charting capabilities leaves a lot to be desired. The dashboarding capabilities leaves a lot to be desired. There are a lot of times, for example, at my last company, they wanted Dynatrace data in addition to a bunch of other stuff dumped into one place. It was not just performance metrics. The CEO wanted his business metrics in the same place as the performance metrics along with a lot of other stuff. However, Dynatrace could not export this type of stuff.
You have about a 5,000 line limit or you have to set up your CSV file just exactly. DC RUM can do it. It might take like an hour sometimes, but DC RUM can do it. I would really like to see the ability to export, in Dynatrace and AppMon, in essentially in a nice format of whatever you want to whatever else. That would be fantastic.
For AppMon, there is always room for improvement: charting, dashboarding, and user management. However, that is pretty much our fault with LDAP. The onboard process itself is a pain, even though we have scripted so much it, it is just very repetitive. There is a lot of alerts and things like that out-of-the-box that do not need to be there or that just do not do the right things.
For Dynatrace, I feel like it just needs a lot more technology support. I know they are trying to essentially get rid of AppMon and move toward the Dynatrace way of doing things. However, we are a multibillion dollar bank. We are not up-to-date. We are not going to be microservices for a long time. We are not going to be container for a long time, and we are probably one the most expensive clients that they have.
We are the ones who are going to drive a lot of the money factor so they need to have that. They need to have integration between the current set of tools so we have the ability to onboard five or six apps, then we'll also put the AppMon agent on it and show people the difference between it. It needs to be better integrated.
All of our team will go to a five minute sales meeting, if they were like, "Look, you can do this with a script." We are sold.
We do not want to do any of the regular AppMon stuff. However, when you have to convince the CTO that we are going to completely rip out the entire monitoring solution which we just spent the last 15 years trying to get a process set up for, and now we are going to redo it. That is not going to go over well. That is not a good conversation.
You need to have that ability to do MQ. I don't care who uses MQ, but apparently we do. If you can't look into those messages, then you just lost a half of our organization which can't be monitored with it.
With stability, I have never run into issues with Dynatrace causing an issue yet.
I have run into AppMon causing issues. There have been a lot times when I have waited for a release. AppMon does the release, then it ends up taking down our application. Now, the fix is immediate, but I have already loss face.
The last time, I took down the main application that lets you call tow trucks. It was just a monitoring loop, a simple thing. They fixed it in a patch. They knew about the issue and they told me immediately what the issue was. I got it fixed in 15 minutes. It took me six months to convince the team to install it in the first place and took me another seven for them to give me another shot at it. It was not a problem that showed up in QA, for whatever reason. So, I could not convince them that it does not exist anymore, because I could not show them any evidence that it existed in the first place. Let alone that I fixed it.
I have never been more confident in a tool because I've seen how easy it is for it to deal with the .NET Applets. That is a big problem in AppMon. Unless every single person is naming their Applets, the exact same way and following the exact same pattern, it becomes an issue. The new tool does not run into that at all. Similarly, you can script it so it just automatically blasts across the organization. As long as it has the PurePath capabilities, somebody who is running, for example, the actual web application that tells people their accounts, that might be a different, more in-depth use case for AppMon versus Dynatrace. So far, the scalability of the solution is phenomenal.
Anytime I can't find the answer immediately in docs or answers, I just open a ticket. They are very good about giving you a response.
Initially, I am talking about two or three years ago, I think they did not have enough personnel staffed there. Therefore, it would take them maybe two or three days to get to your ticket. Now, it is maybe the next day you will have a pretty reasonable answer and that is provided you did exactly what it says in the support ticket. For example, make sure you upload your support archive. Otherwise, you will burn a day and they will send you an email requesting you just upload this. That is shooting yourself in the foot, and that is not their fault.
I have previously used siloed monitoring tools. I have used SiteScope. My current company uses ITCAM. They are okay. They get the job done.
At my previous employment, we used SiteScope and that was quite literally the way that I thought about it in day-to-day life, if you do not really give a crap about it, just put SiteScope on it. However, if you actually need to know if it is working, it needs to have Dynatrace.
I was always pushing for that sort of thing. There is stuff like Wiley where you are not getting 100% monitoring. There is another tool, one is a very new company, and it seemed to get the job done but that was only because we were using Citrix Xenapp. It was specifically able to decode the traffic for Xenapp and XenDesktop, which was what we were looking at. Apart from that, I have never had a situation where I was like, maybe we should not put Dynatrace on this. I have never run into that.
I was not involved in the initial setup at my current company, but I was at a previous company.
For upgrading at my current company, that is in process. We are trying to figure out if it is better to blast it across the organization. We have five Dynatrace servers. They are all completely at capacity. They are all set at large. It is a really big deal for us to try to switch anything over. Right now, we are trying to figure out, do we just upgrade our collectors and hope for the best, or do we do it in QA and then in production?
A lot of people do not run the right thing in QA. It is never the case that their QA is identical to production. So, is that a good indication? I have run into issues before when upgrading from 5.6 to 6.1 expecting that all the bugs were ironed out. That is when I took down that application. Now, I do not have confidence in this upgrade process.
For the Dynatrace Managed version, that setup process was incredibly easy. It took 15 minutes.
Just go with Dynatrace. Just start with Dynatrace. Do not go into AppMon. Start with Dynatrace, because AppMon is going to give you so much extra stuff that 99% of your user base will not need it, including yourself.
You don't need AppMon. I am a hardcore AppMon guy and I am still saying this. It is a lot nicer to be able to start with the Dynatrace solution, be able to script everything, and start integrating the new thing than it is to try to do the old tool set.
I started in the PDP program at Dynatrace. That was when they were still Compuware. Then they became Dynatrace, and I went to a different company, now I am at PNC. I have done the exact same thing for several years in different places.
In my current and previous positions, AI is not important when it comes to IT's ability to scale in the cloud and manage performance problems.
The previous company bought it, and they did not even set it up. I onboarded a bunch of apps. So, they were way too fledgling to try to start looking into it.
What I would use AI for is if it could assist me in saying, "These are your common PurePath patterns." A lot of times in the ending part of an URL, they will have /apps, /data/, then they will put something lie the date or some big custom code. For example, we had one application for tow trucks, which would tell them the URL contained in PurePath, for the actual seven decimal place of the geographic coordinates of that tow truck.
This is not a good way to look at data. If the AI could tell me something about it, just mask it out, or just know this is the same type of data as these other ones and not worry about the extra text part of the piece. That would be the foremost use case for me. After that, I am not sure.
I would need to use it a lot more to maintain my own trust factor in it before I would want to try to tell somebody that is asking me what the problem is. Just immediately saying AI says this. I do not have a high confidence enough factor in it, because I have never really used it.
If my organization had just one solution that could provide real answers, not just data, it would probably put me out of a job. Most of the time, when I get a ticket, they will ask me what the problem is. I will point out the problem, and it is something which is you need to code this better or you messed up these settings. Therefore, as far as helping me not have those mundane sort of tickets where I don't really want to waste my time with people. It is fine for the first few, but after the thirtieth or fortieth person, you tell them that you wrote this very poorly. It is better to just have some tool tell them that this is probably not the best way to do this.
That would be the initial benefit of the one solution. A part from that though, all I am doing is onboarding. The new Dynatrace already takes care of this. So, I am not really sure what my role would be afterwards. Right now, the APM is siloed off from the development teams. If you are going the full Dynatrace route with AI and getting the opportunity of the AIs already going to tell them what the issue is. Then, the APM team does not really need to exist anymore, apart from doing migrations.
Most important criteria when working with a vendor: That initial pairing of sales versus FTS. If I could reach out to them and get answers within a day, or better yet, within an hour. That is one of the best things because a lot of times that initial conversation can get derailed so quickly. You are not going to get more than five or ten minutes to pitch it to your boss. They are always at meetings. For example, my boss, at my previous place, I would be able to sit with him and talk to him about this thing. Then he would get, maybe, five minutes a week of his bosses level. That is the person who is going to sign the paycheck.
Therefore, when he goes to a meeting, and it is a week later, he gives the spiel and has it all ironed out. Then, his boss asks him, "What about this?" Now, he does not know the answer, and I can't get the answer, then I need to get somebody on the phone stat to give him an answer. Otherwise, we have to wait another week. That is a big deal for us to have that communication open.
My primary use case is troubleshooting production issues to see which endpoints are mostly calling and get some logs about them.
It's performing well. We use AppMon, and whenever any production issues occurs, we get a time window from the customer explaining, "This time it is a software failure."
We actually did some training on the AppMon UI, then it almost never failed on us. We could trace back exactly where the issue occurred. We came to the Dynatrace event to understand more about it, and see how we could learn more to perform our troubleshooting better.
It used to take a lot of time to troubleshoot. Now, with Dynatrace, we can actually see the logs anytime we want. I can just find the problem. It has improved performance from a time perspective.
The most valuable feature the solution offers right now is the PurePath. When we see a web request, and something failing, we can drill down using PurePath. Then, on PurePath, we can get to the database SQL statement. So, it is a very cool thing. We have our logs and all the stuff, but this is an area that stands out.
The AI thing is a boost, but it is not in AppMon. It is in Dynatrace. If AppMon could also incorporate it, that would be great.
The main thing is more about the dashboard. Currently, it does not keep the snapshot data. It only keeps it for a very short duration. Because of that reason, we cannot get more reports. If we can somehow option a way to preserve the data and keep it for a longer time, or have that feature, that would be good.
It is always requiring us to update the Dynatrace client. There are some issues with Dynatrace. Many times I thought there was a production night deployment and we would do a smoke test. Therefore, I started Dynatrace, and suddenly it says your client needs to be updated. Then I had to go do that two or three times. So, there are some glitches in it.
In the future, Dynatrace should actually try to resolve those issues with AppMon. The new Dynatrace version, I think will not have these issues, because it will be only browser based. So, there will be no need to install anything on our machines.
I sometimes see issues with the loading of the data. If we give you criteria, like you want to see the last 15 days of data, it has a lot of requests in it. To refresh, it will take a lot of time. Therefore, we actually narrow it down to a particular instance or event.
My main purpose is to troubleshooting issues, so this way I know exactly what time it happened, then I can just narrow it down to that. However, if you want to see a month's data, it keeps on spinning. Here is an improvement which needs to happen, which is the case with all applications or tools. There is a lot of data, and either we have to change the way we are logging or the application needs to be enhanced.
For some reason, we wanted to create a report about what is the maximum used end point or the maximum called end point. Also, we wanted to create a certain dashboard at that time. Myself, with a DevOps engineer, we called the technical support team, and they helped us create that graph. We got it, actually.
We have used production logging. We have our own custom logging system, where we read the logs and transfer to the database. We have a nice UI for it too, but it does not scale well. It can read the log, but it can read only those logs which we added.
Most of the time it works, but it is very slow. If we want to trace back, for example, to find out what happened in the previous week, we cannot get that data.
The DevOps engineer worked extensively on it. I do the troubleshooting.
It saves a lot of time.
It is the best solution that I have seen so far.
AI is like a new feature or benefits, and it is a cool thing. We have not tried it, but I really like to see it working. It is a great program, and it automatically makes a trendline baseline. Whenever something goes wrong, it can send an alert. There is also a web check feature.
Currently, we do not have a baseline. If AI is there, it can see the trend. Based on the trend, it can send notifications. It is also integrated with various platforms, like social platforms, so we can also get alerts from there.
Most important criteria when selecting an APM solution:
We look mainly for how it will scale and what are the features currently available.
I provide this information to the decision-makers.
We use Dynatrace to detect where we can improve performance. We use AppMon to do that. We're just getting started, in the past six months is when we started using Dynatrace, so we don't have definitive results yet, but we have definitely identified problem points that we can work on.
It's given us a lot better insight into how our application performs.
The fact that it can really analyze what parts of the application are slower than others.
I like that the new version has aggregated Waterfall, but I'm told that it's not gonna come to AppMon. I would like that to happen to AppMon.
I would give AppMon an eight. And the reason I don't rate it higher is that the learning curve on the tool is... You have to have some experience with it to be able to figure out where to find things, and the best way to get to things. But once you know that, it's a really good tool.
I would recommend Dynatrace. I don't have any problem recommending it.
I haven't had any problems with it working.
We have a pretty small environment, so we don't have scalability problems.
We had training from them and they're all been knowledgeable. I'm happy with the training that they provided. Their support people are responsive.
We really didn't have anything. We went to an APM because we had a concerted effort to improve application performance. Dynatrace seemed like it was the best solution for the problem. It covered what we wanted and more.
It was complex, but well-documented. Getting AppMon set up has a lot of moving parts, but they're all documented and it went relatively smoothly.
There was SOASTA and one other.
I don't have any personal experience with AI in IT, but I can see that, as environments get more complex, it will definitely help with finding problems.
If we had just one solution that could provide real answers, and not just data, on our ops side it would help because they use that to dig out the root cause of problems.
When we look at a vendor we're mainly looking to see that their product does what we want it to do, that we can analyze performance problems and that it gives us good feedback on where we can improve things.
We are old school, we don't have proper documentation on how our application interacts with the downstream components. We tried CA Wily, that didn't provide a solution. We had CA Wily for three years, but when we deployed AppMon it gave us the complete picture, how the application interacts with downstream components, the application flow.
Based on that, my development team is developing the application automation diagram using the app map. It's like reverse engineering I guess.
It's really helping out. The term we call the mean time to repair, that has improved a lot. Before we had a "command center" type of setup. We had all the dashboards on the screen and five to six people sitting in front of the computer. When there was an issue they used to call a "bridge" meeting and everybody would login to the bridge.
The AppMon solution helped the OCC, the operations guys, to pinpoint one problem area and to call the specific group, instead of opening the bridge for everyone and asking them to join there.
The mean time to repair and the resource utilization time, they are totally reduced deploying APM.
The Transaction Flow diagram and the class and meta level information, those are really key selling points in the automation for AppMon. Also, the meta level instrumentation and the dashboards that most of the people use in our organization.
Most importantly the back-end components. Most of the back-end components that the application connects to; nobody knows how our application interacts with, for example, the DataPower Gateway. But AppMon really provides that information for us. So finding the gaps is the key here.
One thing I'm missing and that is the JMX from the MBeans, it's missing completely.
Also the reporting. We have a load testing team, they completely rely on the reporting for analyzing the data. They should have a template to create a report and they should have something to auto-deliver the report into your email box.
They also need to develop how to capture the JDBC and MBeans metrics. That is something they're lacking. Also integration for the extension to DataPower and MuleSoft Gateway.
The other feature that Dynatrace should have is - from what I see in Dynatrace in our PoC - when you auto-upgrade the agents, the JVM or the application has to be restarted. But if you have something like an "auto-attach" feature, to attach the agent for the running process, it would not require a JVM restart. That would be nicer. That is a killer point.
We haven't seen it that much but it's promising. We haven't had any downtime.
We have a small portion deployed right now, but they're planning to go enterprise-wide, we're thinking about 8,000 to 10,000 agents, so probably at that time we'll know the scalability. But the small one we have right now, it's good. It's doing well. The UI is faster, people like it.
We used tech support one time. We did in Fabric. We have internal cloud, called C3. So C3 and Fabric, we had a little bit of an issue, we called the support guys to fix it so and they were knowledgeable.
I have used HPE Diagnostics, CA Wily. I'm now doing a PoC with Dynatrace and AppDynamics.
With CA Wily, the problem was the complexity of our environment. It's not the tool, we have too many security layers. CA Wily, there's a breaking point. When we have the agents on multiple servers, it's not creating that link because of the security. But Dynatrace, I don't know how they did it, it's really awesome.
We switched because of the Transaction Flow diagram. There is a complexity in CA Wily that it is not able to integrate into our security layer, whereas AppMon can. That's the main thing that's missing. We were looking for that flow diagram. Nobody knows it. The developers call it but they really don't know what the main functionality is that's behind it: How the Apache layer connects to application server, and the application server connects to the different services, and the services to the back. Nobody knows it unless you have proper documentation. Proper documentation is very difficult to get in any organization.
Upper management really liked the Transaction Flow diagram. No matter what, that's the key.
Comparing AppMon to Dynatrace, I like Dynatrace very much because of the ease of use. AppMon is complexity. You need to know more to use the tool. The adaptability comes when you have a nice GUI that is easy to navigate. I saw that in the Dynatrace SaaS model.
When I'm looking at AI, the problem identification and anomaly prediction is important. It's good to know, beforehand, when the problem is going to happen. The anomaly detection is the key area, and part of the AI I think.
If we had a solution that gives you an alert that said, "This is your problem, this is how you're going to solve it," that would be really awesome. Pointing out a problem to a specific group is a key point. That would really help, instead of globally alerting everybody, alerting upper management. If before they know it, you can solve the problem that would be nice.
When looking at vendors, we have a key set of requirements. Among them are container health monitoring, flow diagram. Also extension monitoring the non-Java applications, or non-supported applications, because Dynatrace works mostly on the .NET or Java applications. There are applications out there which are non-Java based like PeopleSoft. At least we can see the interaction with those components, but it would be nice to see what's going on inside those external components. That's what I'm looking in future releases, more support for things like PeopleSoft.
I would rate Dynatrace an eight out of 10 compared to other tools. The amount of data it provides is awesome. Other tools work on a sampling methodology but Dynatrace captures all the RAM sessions that are running. It's more data, but they have the filtering options so I can pick the data I want. Capturing all the data gives me more insight into what's going on, I can compare a bad response to a fast response, that type of thing. Capturing all the data is awesome.
I've worked with HPE Diagnostics, I've worked with CA Wily, now I'm doing a PoC with Dynatrace and AppDynamics. Compared to all these products Dynatrace stood out.
Currently, at my organization, we have DC RUM, Gomez synthetics, and AppMon. We are getting ready to PoC the OneAgent technology. Our use case for OneAgent will be containers monitoring, AWS monitoring, and microservice monitoring. Our organization currently has a digital services team that will be doing a lot of this cutting edge stuff and our architecture is not being fully monitored by the current stack that we have. We also have a tool called ScienceLogic that we use with our Ops bridge and we have Splunk. Therefore, we are really looking for the OneAgent to fill in the gaps that we are not covering.
Out of the gate, I can tell you that just changing the agent model from having to go, "This is a Java, this is a .NET, this is a web, or this is an app," And getting rid of that to having the OneAgent is an improvement. Then, being able to monitor anything, literally anything on the infrastructure, that will be huge for us.
Another thing is the log analytics. We have to work with the Splunk team. We have a very fractured organization, so anytime an application team wants log monitoring, they have to work with the Splunk team. Anytime we want log monitoring we have to work with Splunk team, it is a paperwork process. It is not that they are not nice people, it is just that it takes a really long time, and by then, the problem may be gone.
We really just need the immediate log analytics, because we might not need long-term analytics all the time. Like Splunk, they do awesome data analytics, but sometimes when we are troubleshooting an issue, we just need to look at where the problem is.
I would like better plugin support, because they are constantly asking us to do plugins, saying "Yeah, we can do that. Use this plugin."
Then, the moment something goes wrong with that plugin, I have no way of getting help. They recommend, "Contact the plugin author." You are kidding me? Those guys do not have any obligation to respond. They wrote it, but they do not have to support it.
I do have concerns about stability. I have heard there are gaps being mentioned and I think that we will lose some features. I am not entirely trusting that this is fully-baked yet. I am probably going to go back to my organization and say we really need to do a PoC and we need to get on the train. However, I am not stepping down AppMon or our current DC RUM for another year or two, because this really needs to mature.
I am not worried about scalability. OneAgent is going to more scalable than the clunky old solutions that we just put to bed.
I evaluate technical support on a criteria of:
I have not had very positive experiences in the last year. I found them to be arrogant, rude, not solving my problems, and not interested in solving my problems.
They are very insulted if you ask them questions. They refer you to read the documentation. I am like, "Well, I already read the documentation and that is why I am asking you a question."
"Google that," is not the answer that I want hear out of tech support, because I already googled it. So, I have been rating them very low. It is actually the one pain point I did bring up to my sales engineers and sales support. They are aware.
Our organization just dumped HPOM and the plug was pulled on it in 2017. We are getting ready to sunset SiteScope and HPE BSM. The really siloed monitoring tools are archaic and high maintenance. The infrastructure is too big and moving too fast to be constantly updating old-fashioned tools.
What I really am excited about is that they just announced that it will be doing session replay for the DC RUM part of it. We currently have IBM Tealeaf and we hate it. We hate its guts. Our management hates it because we rolled it out to one application and it cost us over a million dollars to do that. We wanted to roll it out to a second application, and IBM wanted another million dollars. We already have the on-prem and a trained administrator. That was the licensing that they wanted that million for.
Dynatrace just said the session replay that DC RUM will have, it will put IBM Tealeaf out of business. Thank God, because that solution literally is from the 90s. I am not kidding you, it was coded in the 90s, and it is extremely brittle. It is hard to maintain, and it is clunky. If they can get that replay to be smoking, then they will make so much money. All the Tealeaf customers will stampede over here.
We stood up AppMon five years ago. It was not straightforward; it was complex.
We paid for a Dynatrace guardian to be onsite for two years. The guardian rocked. We really loved the guardian. If we had not had the guardian there, we would not have had an adoption rate like we got.
Thursday and Friday last week, I got pulled into what they call a triage team. They had a problem in production where all these people were in a war room with all these fingers being pointed. Nobody had the full picture. I am using AppMon, which performed beautifully and management was super impressed. Just using AppMon 6.5, the visits, being able to drill down, and find the answers there worked. However, if I had had the OneAgent, DC RUM, and SaaS solution altogether, I would not have been in there for two days. It cost our company so much money and they still do not know what the answer is.
This is an expensive solution, but it is also worth the money.
My organization evaluated Nagios and a start-up company. Then we evaluated Dynatrace as well, and it just outperformed the others.
I am really excited about the AI that I am seeing out of the OneAgent because I was just at the IBM conference last year. The IBM AI is still pretty much a toy and I have not really seen the rubber hit the road with their stuff, but the rubber hitting the road is here with the Dynatrace AI. From what I have seen, it will be a key tool set for us just to pin down problems and get answers immediately.
Most important criteria when selecting a vendor: Everything that they are doing is right, except tech support and plugins.
I'm an admin so my primary use case is to manage the product and the software. And whenever there is a user, they have a question, they don't know how to browse or look at the PurePath, that is where I come in, to show them how to.
It has helped us reduce time spent on root cause analysis. Whenever there's an issue, there are multiple teams involved, and we are using multiple tools: Splunk, Nagios, etc. If we have one product that gives one solution to the root cause, the issue, it reduces some of the operation time for our team, so we can spend most of our time on automation or something else.
The PurePaths. The ability to see the transaction flow of the web request. It's very useful.
It's valuable because our developers, when we have an issue, they can drill down to see exactly where in the application, the call in the application; where the high response time is, or where something is wrong. At least they know where to look for the solution.
I want to see an increase in the load, at least to 7,000 or 8,000 transactions per second, just for it to stable for this upcoming holiday, if we do not transition to the new Dynatrace product yet.
What is missing right now is, it doesn't tell me where the problem is. So whenever we find a problem from the customer, or from our incident management team, then we go into the tools and look for it. So we want it to have more problem identification, somehow, just like the new Dynatrace product. They have the problem tab, you can go in there an see it. But AppMon lacks that.
Right now, for AppMon, the maximum handling load, the transactions per minute, is around 6,500. We had an issue on Black Friday or Cyber Monday, some kind of stability issue, that users who could not log in.
It's not very scalable, we cannot add another server onto it. We have a little trouble there, so that's why we're trying to upgrade to this new Dynatrace product to see how it will go.
We just submit a ticket, and it will probably take them one day. That's sufficient for us, for now.
The resolution time is also about one day. Right now, one day is not good enough. We would like it to be in the hours. So that's why, another thing with this new Dynatrace product is they have the messaging system, something like that. And they have the server that would monitor our Dynatrace server, where they can see everything, what the issue is already, so they can resolve it quicker, instead of our submitting a ticket and waiting for one day.
I was not involved in the set up of the app. I remember they were doing it manually. They installed the agent, set up the server manually, on each server. So when I came in, we used Jenkins to write a script to automatically deploy to our server. There is still a lot of manual configuration. It's not straightforward.
When it comes to the nature of digital complexity, and IT's ability to scale in the cloud and manage performance problems, I think moving to the cloud is very agile and provides scalability. Right now, we're using AppMon which doesn't have the AI feature. That's why we're interested in upgrading to the new Dynatrace product, which has the AI feature and scalability.
If we had just one solution that could provide real answers, as opposed to just data, there would be a lot of benefit. If somehow the new Dynatrace product - the AI engine that tells you where the problem is - would tell us where the problem is and bring it down to the pinpoint, to the root cause, so we don't have to spend time looking into it, or the developer team doesn't have to manually look into it, that would be good because those actions require time.
When we were selecting a new APM solution, the criterion of our company was performance. The performance team, they mainly use this tool to test how load capacity runs. So they are mainly testing for the performance stability of the application.
I would say that if you haven't used AppMon before, then just go and order Dynatrace, instead of going to AppMon and then transitioning.
The one that we recently scripted was just to see if an application was up, it was a very simple script. We had an issue with a vended solution at the university in which the application would just go down and the vendor had said that, "Oh, it never goes down, we have 100% uptime." We didn't have a good way to monitor that before. We said, "Our students our reporting that they cannot log in." With Synthetics, we wrote a very simple script that went to the landing page and tried to type and hit "enter," and that's it.
We have two-factor and some other things that prevent us from going too far into it, and we haven't figured out a technical solution for it, but it's a very simple use case of just making sure that the application is up.
Dynatrace itself is performing well.
We now know when the application is down, as opposed to students opening tickets and letting us know. So it's more proactive.
That it's always running.
From what I've learned today, here at the Perform 2018 conference, there are two things that I really want to see.
Number one, the thing that is preventing MSU from moving forward with Dynatrace right now is that we can't tag our customer traffic with a customizable tag. All of our students have a unique identifier and in AppMon we tag that and we can search by it very easily and it's very useful. But in Dynatrace, you can't yet customize and find people like that, so that's really preventing us. I heard that it's being worked on but I'm not sure when it's coming out.
The second thing that I'd really love to see is - I'm very impressed by all the new features that I've learned about at the conference - and one of the features is "impacted users." I would like to see a rate of impacted users. For example, how long has the problem been going on: 100 users in five minutes. Does that mean that in 3 hours if we don't get this solved, we're impacting x number of people? Understanding the rate at which the problem is impacting people would be a cool feature.
It seems very stable.
With the AI, it seems very scalable.
We're going through a proof of concept right now, so we work very closely. They're knowledgeable and we get the right person whenever we call.
We didn't have a previous APM solution. We didn't know we needed this solution until we saw it. It was literally students calling in with problem tickets.
I'm just a supervisor, so I sat with our technician during setup. I didn't do any of the actual work, but it seems seamless. It installed in about two minutes. I really wanted to see it. I have to go to the assistant director and, eventually, the director and try to explain why I think we need the new technology.
They went through a whole eight month process. I wasn't there, I'm a new manager, but I understand they had over a dozen companies. They had spreadsheets of all the pros and cons. They went with Dynatrace because it just has more features.
Regarding AI and the ability of IT to scale into the cloud and manage performance problems, we don't have the new technology yet, we don't have the new AI, we have the old AppMon and Synthetics. But it seems like it's very important.
We have used siloed monitoring tools in the past. The university is very old and very segmented and different departments use different tools and we don't all talk to each other. We still have this problem today and we're trying to get more user adoption for Dynatrace. But it's difficult.
If we had just one solution that could provide real answers, and not just data, I think that we would spend less time working on things that probably don't matter, like mundane routine operations.
To me, the most important criteria when working with a vendor is responsiveness to issues.
My advice to colleagues would be, do your homework. But, I would be surprised if anybody would beat Dynatrace.
When we first got Dynatrace, it started just as a developer's tool. As we used it more and more, we started using it not only for applications but for hardware. Now we're using it for servers. it's evolved over time.
The use cases are monitoring, alerting, triaging, and for statistics and dashboards.
It's basically used to make sure that everything is running okay. You can check dashboards in each group. The DevOps group can use the dashboard and say, "This is how my code is working." So it's more peace of mind, and to keep things available, which is really important to us: availability and reliability. That's what it's being used for.
At product start-up, it was not very visible at first. But now, since people have actually been using the product, and they've actually seen the visibility and what it can provide, it's doing well. So that's why we're deciding to upgrade.
We've had the product for about eight years now. They started using it more and more, so the company decided to invest in it more and more.
Right now in an AppMon, when you have DevOps teams looking to see where transactions are going through from business transactions, and they want to set up measurements to see how things are. And dashboards - dashboards are extremely important now. Thresholds are important.
After today, at the Perform 2018 conference, I can say more because I can see, even though we're AppMon 6.5 now, and we're going to Dynatrace, I was really impressed with everything I saw. The fact that I was impressed, and I was seeing everything, what the future looks like, made us feel like we made the right choice as to what's going on.
Dashboards is one, troubleshooting is another. I come from the monitoring perspective, so the ability to triage quickly is important, and the ability to alert and tell people where the problem is, that's what I really like about the product.
I haven't had a chance to go through all of it, but I would like to see the ability, from an administrative standpoint, for it to collect statistics. I want to be able to see the servers that the agents are installed on. I want it to be able to start doing collections for me by platform: How many Linux servers do I have? How many Windows servers do I have? Statistically give me the information of how things are performing, but I want that in a dashboard, where I can look at a dashboard and I can look at a section. So the ability for me to drill down will make it easier for me.
Ours is pretty stable, being that we have AppMon. We have collectors and DTs in there, they're load-balanced all over the place. And every now and then we have issues, but it's been pretty stable. It has to be, because it's being relied upon more and more. If it was not stable then management wouldn't say, "Okay, we're going to go this direction."
Stability is extremely important because when you're in my business, you need to have data available 24/7. If I have systems that take a hit, and I can't depend on my dashboard to pull up data, then that's a problem. And if they can't triage, then that's a problem.
We have encountered downtime. When we have downtime we look at it and say, "How quickly can you get it up?" The good thing is that when we had downtime, we also had load-balancing in place. If you have that in place that really helps out, because then you go from one to another. It doesn't mean that it wasn't down, but it wasn't down long.
We can't afford to be down, because being down costs money. Everything is equated to money, to some degree. If this tool is not available during this time, how much money am I losing because I can't triage it? How many resources am I using? We've had some downtime, but not much. Even when you do maintenance, it's done a certain way.
Nothing is 100%, but it's pretty reliable. It had to be impressive enough for our management to allow us to invest into it going forward.
It scales pretty well. I'm from a pretty big company. The solution started out small and scaled out, because we have a lot of collectors.
We scaled up. The good thing is, that's where Dynatrace SaaS is going to come in, when you look at what it's costing us to maintain the hardware on-prem, and do the analysis.
Scalability: We're always building more applications. And the fact that we can get rid of a lot of hardware and run that tool is really important.
Scalability is really important with growth. You need something that can scale quickly and not have to go through this whole process. It's one thing to look at it and do an analysis and say, "Okay, I need to scale." But, it would be easier if you could look at something and put in a formula and say, "Alright, we can scale based on this." Or something that could tell us. "You're at a point where you need to scale," so that would help out.
It's easier when you can get to a level-2. We've had guardians on site for a while. Because we've had guardians on site they made it easier for us to get to support. So we know how to navigate support.
From my perspective, support has been quite satisfactory, because it's not like we're a novice going at this. We understand, these are the people you need to talk to. And they recognize the urgency when you're calling. So, I feel comfortable with saying support has been pretty good for us.
I wasn't involved in the initial setup, but since then we've upgraded so many times. The upgrade is pretty easy. Complexity comes in when you have to schedule things. If I have a DevOps team that's in prod, once we were able to actually cookie cut the upgrade process, it was easier.
It also depends on the different platforms you're dealing with. You're dealing with Windows versus Linux, and you have to inject code into certain places, that's an issue.
I'm glad when Dynatrace says they have a different agent, because I wasn't crazy about the agent, the way we were deploying it. Not only me, but you have people who are quite protective of their code, and if you've got to inject an agent, they're not really comfortable with that. But with the new agent, that makes it easier for me too. That's another reason I was happy, that's one of the things I asked about: How are we doing the agent? Even though we were automated, there were still some configurations. But with this new product, it seems like deploying agents is going to be awesome.
I'm excited because I like the AI piece. I want to get rid of the thresholds piece. That's part of the excitement as to where we are going to go, because a lot of the conversation occurs when you have to say, "What's the real threshold based on the applications that we have?" We have over 500 applications, easily. Each one has its own behavior. Then everyone wants to discuss a threshold. Now, I'm thinking, with Dynatrace SaaS, we don't have to do that anymore. Bit it will be awhile before we get there.
AI is really important. We're on-prem right now for hybrid to go to the cloud. So when you have the numbers and trends based on what AI can do, it helps you along the way. I can run reports and see the things that I need that will satisfy my customer, that will make it easier for me. We are looking towards the cloud, we do have applications up there. It becomes a sizing issue, as far as what do you need, because you have to pay to be in the cloud. So AI is really important because it will not only help us troubleshoot, it will actually predict some of the things we need to help us get there. I can't put a number on the importance of AI right now.
It's going to change a lot of analysis and things that you have to do. The one thing I would like to see is what's involved in setting up the AI. Because I'm excited about it, I want to see what's involved in setting it up, and see that component.
Regarding siloed monitoring tools, portability is one of the challenges. Siloed monitoring tools make it really hard to port out to other places. You get data, but you can't necessarily use it as far as fitting it into other areas. That's the problem with siloed monitoring tools. If you're good for that scope then you're okay, but when it's time to go beyond that scope... Going enterprise-wide is important. So, where siloed monitoring tools were okay at the time, they're just not keeping up what's going on now. The monitoring field is evolving. It's more than about just monitoring. Monitoring is one thing, alerting is another thing. But the AI part puts it together, makes it easier, not only for you to do your job, but the self-healing piece that goes with it. That's really exciting, when things can self-heal and then just report on it. That makes life a lot easier.
If we had just one solution that could provide real answers, and not just data, the benefit would depend on how you look at a benefit. First of all, we would be more efficient. However, I'm not sure if it would help from a resource standpoint, because then I'm not going to need all those resources, if I have a tool that's going to do everything I need. But, it would help in the sense that, if I can resolve something like that, and things fall into place. That would greatly help. And not only that, then I could just depend one tool. Right now, people look at a solution as, "I need a tool for this, and a tool for that." But if I can move towards one tool that will provide me everything I need, then that would be great.
Our most important criteria when working or selecting a vendor: Of course, there's always cost. Stability is another. The amount of time they have been in the market. Then we do a PoC and see if they can meet the use cases that we have. We have a standard, some 125 use cases that we put out there. We see they perform based on those use cases. It's completely agnostic. I don't care who the vendor is. Can they meet my use cases? Do they have stability? Do they have the reputation? And then we negotiate cost.
Right now, the solution we have in place, I would give it about an eight out of 10. Things are changing, technology needs to change. It's been doing a job, but I mentioned the things I asked for. I'm looking for more, so that's probably why an eight. It's doing the job, but I'm looking for it to go further.
They have a new product that's going to go further, so that's how we do it. Now, the new solution, I'm not sure how we're rating it yet. I have to see what it's going to do.
But would I recommend it? Yes. Because we have the experience, it has done this and this for me. Can I depend on it? Yes.
If a colleague at another company was looking to implement a similar solution I would need to know they were looking for before I could advise them. I would tell them from my perspective, this is what the tool has done for us. If they ask me the questions, I'll tell them exactly what I think. At this point, would you recommend this solution? Sure, because it does this for me. If you're looking into looking at code, at that part, yes it does that for me. If you're looking for monitoring, yes it does that for me. I would recommend it.
We currently use AppMon. It has been performing great compared to what we replaced.
The main use case is application performance monitoring and availability.
The main benefits will be more proactive, high-quality performance monitoring and availability with more insight into the application. It is very impressive.
It has helped my business move forward probably just by taking the next step that everybody needs to think about, which is AI and the possibility of cloud. I do not know what direction the company will go with that yet.
We are still struggling a bit with finding an answer quickly. It is all there, but there is so much that it is still really hard to figure out exactly which way you want to go:
It is a little overwhelming with the amount of data that we have. I think it is more of an issue on our end. We just need to get more familiar with it. We just started implementing it last June, are still relatively new, and are currently struggling with some performance issues.
We have already seen the next release. It is doing so much more than what we ever expected.
Right now, we are struggling a bit with stability.
We have outages. We have not pinpointed what the problem is. We have data loss. We have had to restart and have not narrowed down exactly what the problem is yet. So, that is in the works.
Scalability is good.
Because we are having stability issues right now, it depends on what we need to do:
So, this is a little uncertain at the moment.
Technical support is good.
There was only one time that I had to contact someone at two o'clock in the morning and the response was not as quick as I thought it was going to be. However, we addressed that with meetings. I think they have done a lot better job of their support.
We have used siloed monitoring tools, specifically CA APM. The challenges were probably the end-to-end and getting the full picture of what the problem was. There always seemed to be some major piece missing and Dynatrace has allowed us to get as close to that big picture as we possibly can. Then, with OneAgent, it has to provide the problem to you. It does not really get any better than that.
We picked Dynatrace because of the value that it provides.
Our environment is very complicated anyway, so the initial setup was a bit of a struggle, but only because we have so many applications and JVMs that we have been working on for long time. We had to move everything from CA APM to Dynatrace, so it was a big conversion process.
We found a tool that can be utilized by testing, DevOps, marketing, software engineering, and monitoring. Before, we always had everybody doing their own thing. Now, everybody's utilizing one tool, which is huge. That is a huge savings.
Definitely look at Dynatrace if you are looking to purchase an APM solution.
We did evaluate other solutions, but I was not a part of that process.
If I had just one solution which could provide real answers as opposed to just data, the benefit to my team would be time savings. We are scrambling all the time to try to figure out when there is a problem. If we could have something else telling us what the problem is, we could spend more time fixing it, providing valuable dashboards. and other valuable monitoring, then have a better proactive monitoring environment. So, it is huge.
The role of AI when it comes to IT's ability to scale in the Cloud and manage performance problems is huge and really important, especially after seeing OneAgent. We will probably be moving to that, then probably be upgrading to version 7. I think that is the direction that the company will be going. They do not say that they are not supporting it, but it is highly encouraged that you jump onboard the OneAgent train.
Most important criteria when selecting an APM solution: Something that was less overhead for us. We were finding with our old tool, which we still support because we are still not off of it, to get the same thing that we get with Dynatrace required more servers and effort on our part to address everything that we wanted to from the performance side. So, there is less infrastructure to support and it is a little more consolidated.
We are using AppMon UVM and Dynatrace synthetics, which we recently implemented. We are looking forward to using the new releases of Dynatrace in the future, where the use cases then will look at the following:
Being able to identify quickly what the problem is. It allows you to better utilize and focus your talent, and tell management that you are not only looking at problems, but how you can ultimately make the product better for the end user.
Business Case: In our implementation, AppMon and UVM are fairly infantile. The implementation was less than six months ago which means that we have been able to scale down some of our infrastructure. Not needing as much infrastructure to run compared to what we have been running in the past. From this perspective, it has been cost effective in allowing us to scale down infrastructure.
It allows us to make better business decisions based on what we are seeing, not what our customers were telling us was going on, but what we are actually seeing. Using this, I have been able to identify what the customers are experiencing. It allows us to optimize our site, our eCommerce solution, to better suit the needs of our customers. We can actually see what they are seeing and almost feel what they are feeling, because of the eyes that we have into their experience. It allows us to better code and better develop, allowing us to be able to optimize our website based off what we are seeing the customers truly experiencing.
The ease of use. Being able to readily identify a problem and be able to get someone there to fix or manage it properly and quickly. Our eCommerce website deals with users and time is critical. One of the best things, we are able to identify problems quickly and are able to resolve them.
They have had years developing this technology. When we go through it and we use it on a day-to-day basis, we see some things and think, "Hey, if they just had this, man this application would be a lot better." Then, in the next release, it comes out and it happened. We are using AppMon 7.0.15 right now, so AppMon 7.1 is coming out. Everything that I have identified in the version that I have as needing to be improved/fixed has already been addressed in the newer versions. Therefore, I can't think of anything that they have not addressed.
One of the things that find to be a challenge is I tell everybody that you almost have to be a private investigator to try to figure out what it is you want and how to get it. What I would have wanted was a solution that makes that process less cumbersome. 7.1 has already identified that. The Dynatrace solution has already identified that because of the way that PurePaths were looked at in Dynatrace compared to what they were looked at in AppMon. That would have been the one thing that I would have said, "If this product could be better, it would be from that perspective," but they have already identified it. They have already come up with the solution because I was not the only one that thought of that. Other people have thought of that using it, and they said, "Okay, this was a gap in Dynatrace, that gap has already been closed." That was the one thing that I had and they have already identified it.
We have not had any stability issues with it at all. This has been the most stable solution that I have worked with. I have been in IT for over 22 years and the solution that we have implemented here in the last six months has been the most stable solution that I have worked with in all my 22 years in IT.
It is limitless when it comes down to being able to scale up or even scale back, if we need it to.
I have definitely had to use technical support. The technical support team, customer service team, and Dynatrace altogether have been the best team that I have worked with industry-wide. Whenever I have a problem, most of the time they are hounding me for more feedback on the problem than me having to hound them. They have been great. I have a monthly call with technical resources, and I also have sales resources on a monthly call. Whatever our need is, they are there to help us, hold our hand, and walk us through it. It has been amazing.
We have used siloed monitoring tools in the past. I will not mention any particular names, but the challenges have been that they are silos. It does not give you a holistic view of everything that is going on in your application as compared to a solution like Dynatrace which allows you to see from start to finish. With Dynatrace, you get a complete holistic view of the application, and it helps you not point fingers, but be able to identify visible problems.
I was hands on in the setup of the solution. Initially, it seemed a little daunting. Once we started working with it, it was a very easy solution to implement and put in place.
The vendor that we brought in to help us move our platform from Legacy to NextGen was using Dynatrace in their dev testing. So, just for amount of consistency, we continue to use Dynatrace, and we can see why they chose Dynatrace to use as their dev testing. It was the best tool out there. I do not know what tools they have considered in the past, but I know that was the one that they brought to the table and it has proved to be a valuable tool for us.
I would recommend to give Dynatrace a call if you are looking for an APM solution.
AI is the solution of the future. I say the future but actually right now moving into the future. It is interesting looking at some of the applications using AI to solve problems, like betas, and being able to integrate that with Alexa for things like adding voice control to identify and solve problems. I think going forward it will be the way that IT works.
It would be invaluable to have one solution which would allow you not only to gather data, but to be able to make intelligent solutions based off of that data. It would be industry changing.
Most important criteria when selecting an APM solution: Cost is one of them, but we also needed something that did not just collect data. We needed something that provided a true benefit. If I am going to spend my day focusing on what a user is doing and what is that user's experience like on my website, then I need a application that can go and figure out the things that I am not seeing and what I am missing, so the tool that I am using provides the information right now. What I have seen from Dynatrace is that it takes that to a totally different level.
It is performing well. It makes our lives easier as we drill down to problems. If something happens in production, and a client reports the issue, say our page is loading in 10 seconds, and the issue comes to the performance testing team, we can get the exact API call or query level regarding what is causing the slowness. Then, our job is to open Dynatrace and test the filter on the time span if it is there. Some matters cannot be captured, because by default, Dynatrace does not capture everything. For some matters, we need to add sensors, then handle it.
Whenever a developer does a code change or new feature implementation, it comes to performance shifting. We just do not raise a bug saying that your page is loading in 10 seconds, please go and fix. What we do is analyze Dynatrace and test PurePath. We analyze metal hot spots, which are the exceptions, and we share data that in AppMon where we have a facility where we save station and take screenshots of what metal or what class it is causing issues, and share that to the engineering team. With the bugs, we have to go and do line numbers, and the same thing with the code.
Getting the EM data, we have to open a browser. Generally, one of the asks from our clients or our engineering team is to change this. However, there is not a performance single tool, which opens a browser. That is one of the problems.
It is very stable.
For our company, it is on-premise. It will be around a 100 plus hosts and 3 different servers, so it is fine in terms of scalability.
I have not used technical support. There is a strong user community. There is no need to talk to the technical support, because all the questions which I have had, all the solutions were in the documentation. Or, I have been able to post a question to the user community and get an answer within a day or two.
No. This is our first APM solution. We have not previously used siloed monitoring tools either.
I was not involved in the initial setup, though I was involved with upgrades.
We take care of our own environment. We are responsible for our performance setting environment. We place the first sensor. We do all those things.
I would definitely recommend this solution, because it has all of the new features coming with Dynatrace OneAgent. It will be awesome for any client or company.
If I had just one solution which could provide real answers, not just data, and benefit my team, it would be one solution where you can get all the data stating this is the root cause and this is the solution. That would be awesome.
AI will definitely play a bigger role when it comes to IT's ability to scale in the cloud and manage performance problems. ZSA will need to predict things like rolling insights to the client on the sales and marketing front.
Most important criteria when selecting a vendor:
We are actually using it in the test environment, not in production. That is our use case. What we want to do is we want to test it and find issues either with our core or infrastructure before it goes live.
We are able to fix issues rather quickly, by identifying then fixing them. Therefore, the efficiency of the organization has improved. We are spending less time fixing issues.
Business case: We had a problem with the previous solution where we were spending more time on it. That meant increased costs. Now, we are spending less time, so costs have decreased. That is how you improve the efficiency of the organization.
The ability to use PurePath in analytics is definitely the most valuable feature. It helps you pinpoint issues, then develop and focus them in the right way. It helps to fix them rather quickly.
I am more interested in the end-to-end performance system, not only the server site from the client's side.
Right now, we use the Dynatrace library to build. Then, we would like to generate the overall transaction and the performance it has when the request is coming from different networks across different regions, different periods, and different parts of the world, like the synthetic monitoring solution. So, there are desperate systems.
What we would like to see is to make it easy, more automated with an easier way of doing these things. OneAgent is an easier approach which you just install it automatically into the installed application and it sends you the information. For the mobile clients, we want to get the end-to-end thing that makes it easier to set up, that will be cool.
With stability, we have seen some issues. Then, we fixed them. So, it is stable now.
The scalability is another objective. We are trying to test the scalability of Dynatrace. We have seen some scalability issues because of the non-optimal users of the resources and the infrastructure, so we made it more scalable. So, it is acceptably scalable.
The role of AI when it comes to IT's ability to scale in the cloud and manage performance is very important. The AI definitely works.
Technical support is pretty good. Not only the technical support, but the upper managers, they are also very technically, savvy people. Most of them, not all of them are. Also, some of the sales accounts, they are able to find you the right technical people, if necessary.
In my organization, I am dealing with more than 20 applications. I have to scale with limited resources, so I needed a sophisticated tool to tell me where the problem is. Traditionally, I would be telling you that you have a problem. Now, I am able to tell you that you have a problem right now and what it is. That has made a big difference. I was looking for tool to do this, then I found Dynatrace.
The AppMon 6.5 is problematic in configuring. It is little finicky. When we configured the JVM, it did not work.
Sometimes, people are making mistakes, typos, or even if you just sit down you need to be in the light, either in the beginning or in the end. So, this is always a little problematic for us. The issue is not super small, but it is okay.
We are buying more licenses, because we are seeing more value.
I would advise a colleague or friend to use Dynatrace.
The most important criteria when selecting a vendor:
We will actually be integrating with Mouseflow, which does the same thing. It is not a big deal to integrate and disintegrate it. If session replay does what Mouseflow is doing, then it is probably our best bet.
It gives more visibility into all the coding (the black screen). It gives a nice screen. You can see ups and downs. You can see where the traffic is getting impacted, more on the convergence side.
It is a platform that is very well-suited for marketers, but also for technology people. That is the key: the dashboard.
We are waiting on the new features to see how they perform.
So far, it seems good. We are not seeing issues. We are happy with it, and we have in-house based Dynatrace. Now, we are moving to the SaaS-based. We cannot comment much on the SaaS yet, because it is moving, and we are not there yet.
In the context of adding monitoring to all your instances, we have scaled up, because it is a cloud-based solution. It just scales on its own.
What you need scalability for is around the ease of adding a new platform or new servers to the platform, and that is pretty fast.
I have not spent much time with the technical support. I am an enterprise architect, so the time spent from the technical side is more the DevOps team, and so not much me.
We did use siloed monitoring tools. We used AppDynamics. Our challenges using siloed monitoring revolved around integration. You needed more manpower to manage things. Now, it is just a few clicks, and it scales up. That is the difference.
We switched to Dynatrace based on feedback from the business and the technical team. It was the usability, and how you transform data into a more meaningful presentation or graphical interface.