Monitoring stories

improving IT service quality

These days who has time to check out new software? You could easily spend half a day figuring things out reading the website. Well, we've got you covered with our 2 minute video that explains it simply: what Alyvix is and what Alyvix does. In fact it takes less time to watch than it does to read this article!

Just like private enterprises, government agencies operate web services running vital services that need to be monitored for availability and responsiveness. Today we'll talk with Gabriele Cecco, who created a visual monitoring test case for a regional government. We'll get his insights into both some important features, as well as how to diagnose what's going on when you're building a test case and it doesn't work the way you expected.

If you've worked with Alyvix for a few years, perhaps you've noticed every so often that your test cases stop working for no apparent reason. If the underlying problem isn't an actual system fault (congratulations, your monitoring is working as intended!) then the cause is almost always a change in the interface that you're monitoring.

While some large "breaking" changes will obviously require you to create a new Alyvix test case, more often it's just a minor change, for instance Alyvix can't find a button that's been moved due to a software update, or a multi-user system has persistent window properties. In this best practices blog, I'll show you how you can build more robust test cases so that these minor interface changes won't interrupt your monitoring and keep you from rebuilding your test cases.

Monitoring is essential to keeping IT systems running smoothly. Alyvix Server's visual monitoring approach complements typical monitoring systems by directly measuring what users experience. You can explore these measurements graphically to find and certify severe latencies and service interruptions, potentially resolve them and even prevent them from occurring. That's the Alyvix Value.

End user experience monitoring continuously tests the performance of business-critical applications from the perspective of end users. It quickly tells you about any degradation in performance, responsiveness, or availability that users may experience. This can help you avoid significant problems that can lead to poor customer satisfaction, lost revenue, and negative brand impact.

Real User Monitoring and Visual Monitoring are both user-centric strategies to ensure that quality metrics important to users are maintained. Their underlying methodology, however, is different, leading to a separate set of advantages and disadvantages for each approach. The proactive nature of visual monitoring can help you discover and remedy problems before users even notice them.