To be serious about monitoring you better have an application running on a cloud platform, right ? Public or private, whatever that is, Google, Amazon, Azure or on a internal network powered by OpenStack, VMware ... preferable deployed over a large number of nodes, using large memory configurations and offering advanced dashboards to keep track of hundreds of metrics collected from machines, storage, network devices, applications all available and ready to be ... read and digested by anyone.
The duty admin is confused, what dashboards will be needed for basic performance monitoring and SLA and what metrics these dashboards should include ?
We took a simpler approach with Kronometrix, which can be deployed on a cloud computing platform if really needed, which was designed to be:
- easy to install, manage and administer, aiming for zero-administration
- ready for computer performance analysis, including essential performance metrics
- self maintained and automated, majority of tasks are pre-configured and already set
- simple to read and understand using clear UI dashboards
- available for operation people, ready for large size screens, example 51"
|Operational Dashboard 5 and 30 minutes, zoom in|
- top essential performance metrics, per host, included
- clear and simple UI, no confusing labels, charts, extra information
- two time ranges, allowing operators, sysadmins to check current and past activities
- possibility to drill and zoom in, using the time series data charts
- direct link to console events, alerts and thresholds
- designed for large size screens, eq. 51", on night and day mode