The solution is primarily used for user happiness monitoring. Basically, we look at how well we'll use it to run some arbitrary metrics over the website's behavior. I'm trying to understand how the user experience is going. Therefore, we're doing user experience monitoring. We're also using it for monitoring and listening, where we fire off specific test cases against that site with it.
Typically, it does quite a lot. We can also do health monitoring of the actual service hosts, and servers, and dependencies.
Dynatrace provides us with the ability to actually map out the whole ecosystem and our websites and services that exist within it. You have that whole picture even though it's now a distributed network of products and things. We use Dynatrace to just monitor the health of that ecosystem and manage, and identify where the dependencies are. In doing that, we can also look at hotspot monitoring, so that we can determine bottlenecks within our system. We can use it to follow metrics to help us figure out how fast things should occur, to identify slowdowns of speed - or potential slowdowns - which can cause us to have those little mysterious bugs where suddenly the user experience drops out because something three or four levels down is not behaving itself.
Basically, it's a lot of use cases based on the user experience. There's lots of user monitoring. Lots of looking at where they're entering the sites from, where they're exiting the sites from. The behaviors can sometimes help us detect failures in our overall user experience. It's a lot of user experience management that's assisted via AI. We can use AI to develop, identify, establish, buy, and build trends so that we can look forward to purchasing requirements. Ideally, the AI will make it that we can identify where the system is going to fail in the future. We're still working on that side of things, but we're getting there.