We just raised a $30M Series A: Read our story

Nagios XI Competitors and Alternatives

Get our free report covering Zabbix, Paessler AG, Centreon, and other competitors of Nagios XI. Updated: November 2021.
554,382 professionals have used our research since 2012.

Read reviews of Nagios XI competitors and alternatives

Subbarao Punnamaraju
IT Operations Manager at a university with 10,001+ employees
Real User
Top 5
Clear escalation chains mean the right people are alerted, decreasing resource usage and helping with planning

Pros and Cons

  • "Another feature from the technical aspect, the back-end, is the ability to allow individual users or customers to have their own APIs. They're able to make changes using the plugins covered by LogicMonitor. That is a very powerful feature that is more attractive to our techno-savvy customers."
  • "The dashboards can be improved. They are good, but there is a pain point. To show things to management, to explain pain points to other customers, to show them exactly where we can do better, the dashboarding could be better. Dashboards need to show the key things. Nobody is going to go into the ample details of Excel sheets or HTML."

What is our primary use case?

We use it to make sure that proper tuning is done for the existing monitoring.

In addition, our university has a number of schools and each is a customer of the main IT organization that manages and provides support for all the colleges, like the law school, the business school, the medical school, the arts school, etc. The goal, and one of the main use cases that we were planning and thinking about, was to be able to onboard all the devices, all the applications, all the databases, as required by individual schools.

We also wanted them to be able to create their own dashboard, tweak it, manage it, delete from it, and add to it. 

It's deployed as a SaaS model. LogicMonitor is out in the cloud.

How has it helped my organization?

When we were using Nagios and we had alerts but there was only red, yellow, green. Here, the good thing is that you have escalation: level-one, two, three, which are clearly defined, and what action needs to be taken for each level. The clear escalation chain and tuning helps, because we don't want to wake up the director for 80 percent of the cases. That would be ridiculous. But when necessary, the right people should be alerted, especially for the production environment. If something has been "red" or there has been no interaction for half an hour, it's important to know that and to take the necessary actions.

That's a key thing, being a production-operations team member, because I don't want my team to be flooded with all the noise of alerts for something which can be tackled by a specific team. Having escalation chains, so that the alert goes to the right team to look into that and take action, means the prod-ops team doesn't need to even look into it. We don't even need to ticket it. We only keep aware of it through the daily alert dashboards. That has made a big difference in our overall resource planning, because previously we had 400 to 450 daily alerts. By using this feature we cut that down to 150 to 200 which are "candidate alerts" that production-operations needs to take action on. They may require creating a ticket, or calling the right people, or doing some activity that needs intervention or escalation to the next level. We have been able to cut down on our resources. We don't need to have four members actively looking into the dashboard. We can validate things with one or two employees.

LogicMonitor has also helped to consolidate the number of monitoring tools we need. We had some third-party monitoring, four or five things, and they're all consolidated with LogicMonitor. The only exception is IBM Tivoli Workload Scheduler. But what we did was we integrated that via Slack. I'm not really sure why we weren't able to consolidate TWS. The plan is to get rid of TWS, but we could not do so immediately, until there is an alternate route. But apart from that, everything has been consolidated using LogicMonitor.

We were especially able to consolidate third-party cloud monitoring for AWS. There were discussions about how we could also integrate or combine Azure monitoring resources through LogicMonitor. The team has mentioned that it has plug-ins that it can use to combine that. We also had separate backup scheduling software, a tool that had separate monitoring, and that has also been combined with LogicMonitor.

And LogicMonitor has absolutely reduced the number of false positives compared to how many we were getting with other monitoring platforms. At a minimum they have been reduced by 50 percent. The scope of more tuning and going through the learning curve helped to bring it down. Within the first two or three months, we were able to bring the false positives down by 50 percent. That's a big achievement. That is the main reason we initiated this project of getting into LogicMonitor. There have been further talks internally about how we can eliminate them further, and bring it down by 70 percent compared to the false positives we were getting. That's our goal. So far, it has reduced the time we used to spend on them by 50 percent, both offshore and onsite, as we have an offshore team in India that works 24/7. We used to have multiple people in each shift and we have reduced that down to a single person in each shift. That's a big step in the right direction.

What is most valuable?

Tuning is one of the main components. We like to make sure that only the right alerts are escalated, and that alerts are being sent to the right members, as opposed to every alert being broadcast to everybody. The main thing is the escalation chains. We feel that is a very good thing, rather than sending all the information to everybody at each level. Having the ability to make those sorts of changes doesn't require you to do too much, out-of-the-box. You just need to create the basic entities, like who are the different people, who are the contacts, or email groups, and cover the data source and events which should be alerted.

Another feature from the technical aspect, the back-end, is the ability to allow individual users or customers to have their own APIs. They're able to make changes using the plugins covered by LogicMonitor. That is a very powerful feature that is more attractive to our techno-savvy customers.

In terms of basic functionality, from a normal user's perspective, the escalation chains and the tuning part that are embedded in LogicMonitor are the two most important things.

Among my favorite dashboards are the alert dashboards. Being a prod-ops team, we took the out-of-the-box alerts dashboard given by LogicMonitor and we have kept on tweaking it by adding more columns and more data points. The alert dashboard is something which is very key for us as a team. In general, it gives us more in-depth information about uptime, the SLAs, etc. LogicMonitor has done a good job of providing very user-friendly dashboards, out-of-the-box. There are so many things that we are still learning about it, how we can use it better, but the alerts dashboard is my favorite.

The reporting is something which I have explored, to send me an email every day with how many alerts, in particular how many critical alerts, there were. It's a good starting point. The reporting can be sent in both HTML and Excel and is accessible on the dashboard after you log in. These two things are very good. This is the first feature I looked at once we went live, because I want to know things on a day-to-day basis and a weekly basis. I activated the email feature because I want it to send daily, weekly, and monthly reports of my alert dashboard data.

We use LogicMonitor's ability to customize data sources and it's a must, because ours is a very heterogeneous, complex environment. Changing data sources is important for at least some of the deployments. For other organizations, it may not really be required to change the default data sources provided by LogicMonitor. But here, it was important to change them. That's where the capabilities of the embedded APIs really helped us. I'm not part of the team that makes those changes, but I worked actively with the teams that did, and I always got very positive feedback from them on how they would get the right answers from LogicMonitor. They had to make a lot of changes to the data sources, for each customer, and it worked out well.

What needs improvement?

There are a few things that could have been done better with the reporting. It could have a more graphical interface.

The dashboards can be improved. They are good, but there is a pain point. To show things to management, to explain pain points to other customers, to show them exactly where we can do better, the dashboarding could be better. Dashboards need to show the key things. Nobody is going to go into the ample details of Excel sheets or HTML.

Automation can also be improved. 

Finally, while this is a very good tool for monitoring and responding, if there was a way they could do something like PagerDuty or another third-party solution for alerting, integrate both monitoring and alerting, that would be an ideal scenario.

For how long have I used the solution?

I have been using LogicMonitor for close to a year. If I remember correctly, LogicMonitor was implemented in my organization as a replacement for Nagios. I was actively involved in that project right from the beginning of verification through going live. In the initial stages we may not have been actively using it, but we started learning about the tool and how to implement it about a year ago.

What do I think about the stability of the solution?

Overall, the stability has been good. We didn't have any issues during the phase after we set up and went live. 

The performance was also pretty good. We didn't have to wait for a response for any of the attributes on the dashboard or reporting.

LogicMonitor has the ability to alert you if the cloud loses contact with the on-prem collectors. We had a challenge within one or two months of deployment. The problem was the way we were using the collectors. We were actually using our Nagios server as one of the collectors. We were trying to eliminate that server altogether, because it was giving duplicate alerts.

Initially we had a challenge of not getting any alerts when the connection to the collector was lost. Later on we found that there was a routing table or there were some firewall changes that were needed. I would attribute that more the learning curve and what the best practices are.

Since correcting that problem, we haven't had an issue of any collector being down. There's no question about any of the alerting.

What do I think about the scalability of the solution?

The impression we got when we provided information about the number of servers, the number of end-users, and the number of networks that were part of Nagios back then, was that LogicMonitor said they could expand and double that, if things were to grow. There is scalability in that environment to support a big data buffer. So there should not be any problem with scalability.

In terms of DR, discussions are still going on as to what would happen if there were a disaster. 

As a whole, the organization has to use a monitoring tool. It could be Nagios, it could be LogicMonitor. There was a phase in which most of the schools were using both in parallel. But one after another, they are all happy to be using LogicMonitor. Usage-wise now, it's only LogicMonitor. Nagios has been cut down, so nobody is looking for any monitoring system apart from LogicMonitor.

There are some schools that still need to tweak it and tune it, because they have not given it much attention or have not really been required to actively monitor their solutions. We know where the priorities are, which school is the top priority and which schools were using Nagios more actively. But all the major customers that were using Nagios, once we unplugged it, have been happy with the LogicMonitor implementation. There are a few schools which are not actively using any monitoring system. They may get to the stage of actively using it, but, university-wide, everybody is using LogicMonitor. There is no other monitoring tool out there.

How are customer service and technical support?

We have evolved and have kept on making changes, as per the requirement of the customers and one good thing about LogicMonitor is that it has a very good support system. We have had chat sessions with them to ask questions which help each school, and the IT organization as a whole, to evolve a better monitoring and alerting tool.

The way LogicMonitor support responded during our initial setup was amazing. That's something I really enjoyed a lot. They never said something like, "This question should not be asked," or "This question is not a candidate for the chat session." For every question we would get a reasonably quick answer which we would be able to implement right away. They would also log in remotely and help if something was something beyond an individual's capability. That helped to migrate and complete this process in a quicker manner. LogicMonitor has a very highly talented support team that can answer the questions and help the customer right away. It's been wonderful.

I don't see that happening with all vendors. With other organizations, when you submit questions in the chat session, they'll take the request and they'll say, "Okay, we'll get back to you." LogicMonitor — and it's a differentiating factor — is there to provide solutions right away, rather than putting it into their ticketing system and escalating to level-2 and to level-3.

I really don't know if that level of service is only for specific customers, based on the contractual terms and conditions, or if it is the way they do it for everybody. If this is the way they do it for every customer, they should definitely be very proud of the way they are doing it. Their team is there to help support the customer instantly, versus taking their own sweet time.

I would encourage LogicMonitor to continue that same level of expertise, of people being there 24/7 to support customers. That would be a big differentiating factor compared to competitors.

Which solution did I use previously and why did I switch?

The main reason for migrating to LogicMonitor from Nagios was to eliminate the noise of alerts. It may have been because alerts were not properly tuned, but the visibility with Nagios was not complete. It became a bottleneck. 

Only one or two people had active access to tune things. If anything had to be done, there was just one guy who had to do it. We wanted to move towards a self-managed model. LogicMonitor is a solution which can be in that category, once it's deployed and there is a transfer of knowledge to each school.

We want each department to self-manage: manage their own dashboards and create their own reports based on their requirements. If they have a new device coming up, they can spin up a new AWS instance and onboard that, etc. It's the initial phase which is going to be challenging. But once we have the handover call with the individual customer, it's going to be easy, and that was not possible in Nagios.

We also wanted to have a proper escalation chain, which was not present in Nagios. That's something we have made use of in LogicMonitor.

Finally, we switched to use fewer resources and to speed up turnaround.

How was the initial setup?

The initial setup is complex. It's too picky. I'm a hands-on technical guy, although I don't call myself an SME, but I know everything right from networking, servers, databases, firewalls, to clustering, support, and operations. The initial phase is definitely a little bumpy for somebody who's not completely technically savvy. I understand that it's because there are so many features involved, and there are so many ways for onboarding and using the custom APIs, etc. To me, LogicMonitor, looks like too much of a technical-savvy company. There's good and bad in that. It depends on how you look at it.

The automated and agentless discovery, deployment, and configuration are good. We used that a lot initially. They did a good job with that. One thing that could be done is to make the naming conventions — adding different names like the IPs, the DNS lookup — a little better. They could eliminate some of the duplicate entries when you're onboarding it. I saw a lot of duplicate entries, which goes into the licensing. Apart from that, the way they provide a template or a flat file to the system for onboarding is good.

As for monitoring things out-of-the-box, it seemed that our database team spent more time in configuring stuff, whether MySQL or Oracle, etc. Now, LogicMonitor has come up with a very easy way for configuring and monitoring database components out-of-the-box. But that's something which I felt was a little bit of a pain point. I don't know whether it was that our team made it more complicated or LogicMonitor didn't handle it out-of-the-box.

Apart from that, LogicMonitor has done a good job of out-of-the-box monitoring of the basic resources within the servers — memory, CPU, disk configuration, etc. — as well as for HTTP, the web components.

While I wasn't actively involved in the planning for the implementation, I picked up things from the team which was actively involved in planning and implementation. The process was primarily to engage with LogicMonitor. Our team — the product owner and team members — worked together and was in touch with LogicMonitor to gather all the existing features that were available and how we would make use of all that. That was the initial phase during which we got to know the product completely.

We mapped all of the devices which were in Nagios to make sure we onboarded everything that was in Nagios to LogicMonitor.

We had several internal discussions where we told the schools how we were actively engaging with LogicMonitor to make sure that we would go in phases. The initial phase was knowledge-transfer, the second one was to onboard a school, or at least one application, to make sure that it was tested completely and then remove that from Nagios. We took time to make sure that they were getting proper monitoring and proper alerts, out-of-the-box.

While doing that, we found that there were a few things which were not properly configured in LogicMonitor, compared to Nagios. The goal was to improve on Nagios, minimize the false alerts, and have better features for reporting, dashboarding, escalation chains etc.

We had six to seven people actively involved in the process. Two to three were purely technical, and made use of LogicMonitor support very extensively, especially for some of the customized activities like using custom APIs. From the LogicMonitor side, there were two to three members from the front-office who were actively involved, and on the technical side they designated a couple of people whom we could directly contact on a day-to-day basis. We had a daily, separate session with each of our teams, like networking, business, operations, and DevOps, so that each team could ask questions about its pain points and get better information so that we could do things ourselves and, for things that were beyond us, to learn how they could help. We had a month of one-on-one sessions with them, every day, for two or three hours.

When we initially started the engagement with the LogicMonitor team, they came onsite to run a one-week session with all the key stakeholders: the customers, the technical team, and back-end operations team. That was a very useful session that helped kickstart things. At that point, not everybody knew completely how LogicMonitor works and how we could plan to migrate from Nagios to LogicMonitor. What were the things that we could retain? What were the things that we could just ignore? Overall, the exposure to LogicMonitor during that one-week phase, in terms of customer-engagement, was really a great experience for me. We also had the ability to quickly use the chat session online and ask questions.

The implementation team's role and its way of engaging with the customer was amazing. That's something which I really appreciated. That helped me. Once the engagement was over and the contract started, the online support was available. If we had a problem, we could type in our question or our problem right away. The support team would respond and fulfill our requirements. They would fix the problem.

Our deployment took two to three months. That includes the visits by the LogicMonitor to do some knowledge transfer and give hands-on experience to some of the key stakeholders. But during that time, not all places within the university were onboarded. Some schools were not really interested. I don't think they were properly updated. That was something that was more of an internal issue, because we were doing our own "selling" to tell them what the differences are between LogicMonitor and other things. We had to tell them that Nagios was going to be pulled and that they would be completely in the dark if they were not moving to LogicMonitor. So during those three months, there were still quite a few schools which were not migrated to LogicMonitor or didn't onboard all of their resources. But the majority of them were done in three months.

In terms of maintenance, we have three to four people involved. One guy was actively involved in the Nagios implementation and its maintenance. He was part of decommissioning that and completely taking ownership of LogicMonitor's technical aspects. One person is the product owner who interacts with all the stakeholders, the different schools, to make sure that they have their requirements met using LogicMonitor. One is a manager. And there is a person from the business point of view, who provides his pain points, and what they're seeing on a day-to-day basis. So those four people are actively dedicated — I would not call it to maintenance — but to the day-to-day LogicMonitor stuff.

There are the users as well. Each school has its own applications and services that they offer internally. I don't have exact numbers but there are about 20 of them.

What was our ROI?

It allows us to accomplish more with less by minimizing the false alerts.

And by giving the "keys" to the individual owners, it makes things faster.

Also, as I mentioned, we don't need to have as many people in each monitoring shift, in the 24/7 environment. Previously, we had alerts that went to everybody and everybody was up and looking into why we had a given problem. Now that we are splitting the problems into different buckets, we are not tapping into all our resources' time. That's an area where we're saving. As a rough ballpark, we are saving about 50 percent of the resources from an operations perspective.

What's my experience with pricing, setup cost, and licensing?

We have a separate team involved in licensing. I wasn't involved in that.

Which other solutions did I evaluate?

I believe they evaluated two or three other tools, but I was not part of that process.

What other advice do I have?

For the initial phase, rather than having only one or two functional guys participating, it's always good to have one or two technical folks in the discussions. That helps a lot. You don't want surprises if an organization decides to go live with this tool, and then realizes that technical things are not on board with the ideas of the functional team. That's something I can say based on my journey and experience.

Another thing that is important is to keep on having internal conversations; that you value and give importance to everybody. It's good to educate them. Use the help of the LogicMonitor support team for internal question/answer sessions and do anything that will help them feel more comfortable. It's not about two or three members being really happy with this. LogicMonitor is something which can only be successful in automation if all the key teams and team players are on the same page.

The biggest lesson has been how we could make everybody be part of the mission. Previously, monitoring used to be in the hands of one or two, and each of them had a lot of overhead to deal with. But by doing this, we have reduced the complaints from individuals and each stakeholder. They know how they're configured. They know what the escalation chain is, so they're confident. If there is something not working, it's because of the way they have it configured.

By doing this we have minimized the internal noise. We have given everyone the opportunity to know the pain involved in monitoring and what it takes to have a better monitoring system in place, and how each person can contribute and think outside the box. They know how to put into place the right parameters and the right numbers. Previously, 70 or 80 percent of things were escalated internally. There was no involvement of the particular customer. If there was a problem for a team, it was somebody's problem, not their problem. Now, it has all become their problem. This is a very high-level benefit of using tools like LogicMonitor, which involves everybody more.

I would give LogicMonitor an eight out of 10. There are a few things that LogicMonitor is also learning from their experience with the customer. Most of the customers are giving feedback to LogicMonitor for improvements and to make changes. I'm sure that very soon it will be a 10, but at this point in time, from my experience and journey, it's an eight.

Disclosure: IT Central Station contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
Ed Alexander
Senior Systems Administrator at a computer software company with 10,001+ employees
Real User
Top 10
Advanced analytics enable us to look at trending and detect where there are issues or potential issues

Pros and Cons

  • "Among the most valuable features are the queries and reporting that allow us to look at the utilization of resources, at how well the storage is performing, and to report on which resources are being used by which business units. We can track usage across the entire environment, across applications, business units, cost centers, etc."
  • "There is room for improving the creating and managing or modifying of reports. That is still a difficult task to do and requires knowledge beyond the storage itself. I would love to see reporting improved so that we can create reports by dragging and dropping pieces into a report form and publish a report that way."

What is our primary use case?

We're using it for monitoring and managing all of the NetApp storage in our environment, both in the cloud and on-prem. We are expanding it to use it with our virtualized environment and with our in-house storage environments.

How has it helped my organization?

It's allowed us to tag storage with applications and business units, so that when we have a requirement to migrate data for an application, we can quickly identify all of the storage that the application utilizes.

The solution also provides a single tool for cloud-based architecture, such as microservices and containers, as well as on-premise infrastructure. We're not fully utilizing that part of it yet, but we're in the process of starting down that road. It's provided us with the ability to better manage the storage and to report on it to management. We're able to show them, via dashboards and reports, how the storage is being utilized and what the return on investment is.

And when it comes to the solution's ability to quickly inventory our resources, figure out interdependencies across them, and assemble a topology of our environment, it's very powerful. Because it's SaaS, a lot of the upfront work is being done for us. We're spending less time managing the application and more time getting the data we need out of it.

NetApp Cloud Insights advanced analytics for pinpointing problem areas are outstanding. When we have a performance issue for a specific piece of storage, a volume, or an application we can graphically show the business units, the application owners, how the storage is performing relative to the performance that they're seeing on their clients or on their servers. It helps us to pinpoint where issues may be, whether they are with storage, the network, on the clients, or with the application itself.

The advanced analytics have also enabled us to look at trending and detect

  • where issues are
  • where we would have to do load balancing to keep performance from being degraded
  • where we are running out of space in a given data center or storage node.

The advanced analytics help to reduce the time it takes to find performance issues as well because we run reports that look at trending. We can see where performance issues have not yet occurred, but that if we keep going in the direction we're going, we will have performance issues. That way we can proactively take action to ensure that we don't have any problems. We mitigate them before they happen. It's hard to quantify how much time it saves us, because it's very easy to see the problems before they occur, as opposed to dealing with them once they have occurred. It saves the businesses from having issues and keeps the entire environment running more smoothly.

In addition, the solution helped us to right-size workloads. It has shown us where we have issues, where storage growth has exceeded what is manageable. We've then worked with the business units and the application owners to break the data up into more manageable sizes, so that not only do they continue to get the performance that they need, but we're able to manage the storage without being backed into any corners that we can't get out of. Keeping the data broken into smaller pieces allows us to move it around to ensure that they get the performance they need. We're also able to ensure that the applications that require the highest level of performance get it. And those that don't have those high requirements can be put on lower power, with throughput storage that meets their needs, but without consuming horsepower on the fastest storage where they don't need it. Overall, it provides us with better utilization of the infrastructure.

The right-sizing has also reduced costs because we can more intuitively allocate resources and spend money on upgrades only where they're needed. Instead of just throwing more hardware at an environment, we can do it very selectively, based on the needs of the environment, needs that we are able to analyze through the reports that Cloud Insights provides to us. If I had to make an educated guess, it has reduced our costs by 20 to 25 percent.

What is most valuable?

Among the most valuable features are the queries and reporting that allow us to look at the utilization of resources, at how well the storage is performing, and to report on which resources are being used by which business units. We can track usage across the entire environment, across applications, business units, cost centers, etc.

What needs improvement?

There is room for improving the creating and managing or modifying of reports. That is still a difficult task to do and requires knowledge beyond the storage itself. I would love to see reporting improved so that we can create reports by dragging and dropping pieces into a report form and publish a report that way. We do get by with it, but typically it requires engagement with NetApp to get us the reports that we need.

For how long have I used the solution?

We have been using Netapp Cloud Insights for about six months.

What do I think about the stability of the solution?

It's been very stable and very available. I'm not aware of any times that we've encountered an outage where the tools were unavailable to us.

What do I think about the scalability of the solution?

It scales very well. As we've added new parts to the environment, it has been simple to add them to the Cloud Insights environment.

The number of people in our  company who directly utilize it would be about 100. People who benefit from the product and what it produces would be in the thousands; for example, finance. We generate reports automatically in Cloud Insights to show the utilization of all of the storage infrastructure, broken down by the amount of storage that is being utilized by individual applications or individual businesses. All of that is fed into the finance department so that they can track where resources are utilized and who is using them.

How are customer service and technical support?

Their support has been outstanding. They're very responsive when we raise issues. When we request new reports and modifications to existing reports, they're always right there and ready to assist.

Which solution did I use previously and why did I switch?

Prior to using Cloud Insights, we were using OCI, OnCommand Insight, which is the equivalent of Cloud Insights that we ran on our own hardware within our data center. We switched because it's going to provide us with a better ability to manage and monitor cloud-based data, while at the same time allowing us to do everything else that OCI used to do for us.

How was the initial setup?

The initial setup was fairly straightforward. It was not that difficult because NetApp worked with us throughout the deployment. Since we already had OCI on-prem, migrating just involved taking that OCI environment and recreating it in Cloud Insights. It was not difficult at all. We planned that all out with NetApp and they guided us through it. In fact, they did a lot of the work for us.

Our overall experience with NetApp support during the deployment was very positive. We didn't have any issues that weren't quickly resolved.

The deployment took about 30 days.

Our team has six people, all of whom are storage engineers, but none of us is dedicated, full-time, to administering or managing Cloud Insights. We're more users than maintainers, because it doesn't require much maintenance.

What was our ROI?

It's hard to quantify ROI, but it is definitely worth the investment. What it provides to us has extreme value, helping us to get the full utilization out of our investment in storage. Cloud Insights has helped us to optimize spend by removing inefficiencies and identifying abandoned resources on-prem. Eventually we will be getting there with our cloud infrastructure as well.

We've used it to find storage that is no longer being used by any clients; storage that had been orphaned. We have been able to work with the business centers to determine if the data can be deleted, or should be archived, or moved to less expensive storage if they need to keep it online.

The optimization hasn't reduced our cloud costs yet, but it will eventually, as we move deeper into the cloud. Right now, the lion's share of our data is still on-prem, but we are starting more migrations into the cloud. Cloud Insights will be integral to being able to manage that space and the amount that that space costs us.

Which other solutions did I evaluate?

We were using OCI, which was the predecessor to Cloud Insights. We've also utilized other monitoring and reporting tools, such as Nagios and Splunk. But Cloud Insights gives us a level of information we can't get from the other tools. It's far more powerful.

What other advice do I have?

My advice would be to take a look at it, do a PoC, and see all the value that it can add, and all the power that it provides.

We rely on it to manage the storage, to look at trending, and determine where we're going to have problems. We can deal with things before they become problems. That's probably the most powerful thing that we get out of it. I don't think a day goes by that we don't utilize it.

Disclosure: IT Central Station contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
Mamadou-Diallo
IT Analyst at La Corporation D'urgences-Santé
Real User
Top 5
Helps me detect where the problem is quickly

Pros and Cons

  • "Centreon helps me detect where the problem is quickly. When we resolve a problem quickly, this lowers our overall costs."
  • "Opening a ticket on the website of Centreon can be difficult for my colleague, but not for me because my English is good. However, my colleague doesn't speak English well, as our company is in Quebec and our first language is French."

What is our primary use case?

We use it primarily for monitoring our devices. We also use it for discovery. For example, if a server is down, the first step to take is just go to Centreon and have a look.

I installed this solution on my company's local server. It is on the LAN network. There is a plugin for Centreon for connecting and providing visibility to the host.

How has it helped my organization?

It provides visibility into the network. Without visibility, it is difficult to know what the first step to take is when there is a problem. With Centreon, I can just go to their website if the server is down and it gives me visibility.

We have seen increased usage. That has expanded.

Centreon helps align IT operations with business objectives. It helps our guys in support monitor all our services. It gives you visibility for all services on the host via a link.

What is most valuable?

You have one single view, which is good. As the administrator, the permissions are the same for every account in the Active Directory. 

We use the solution’s Anomaly Detection feature. It helps our teams to predict potential issues. I like this option for its graph. The graph is very good for when you want to look at evaluating your system over time. It also automates the alerting process. 

It is very easy to manage because the visibility is good.

What needs improvement?

In the next year or two, I need to implement a load balancing system because I am using Veeam Backup, which I don't like. 

For how long have I used the solution?

We have been using Centreon for the last couple of months.

What do I think about the stability of the solution?

The stability is good because the operating system is Linux, and we are using the latest version.

We use it with vBackup, and the server backs up to vBackup every two to three days.

What do I think about the scalability of the solution?

The solution is monitoring 6,000 services.

35 percent of the technical support use the solution.

How are customer service and technical support?

The support is good when I contact them directly. Opening a ticket on the website of Centreon can be difficult for my colleague, but not for me because my English is good. However, my colleague doesn't speak English well, as our company is in Quebec and our first language is French. We do have someone there, who helps us with French when needed.

The support checks the command line of our Linux system, which is good. 

You can go to their website to investigate problems. This provides us flexibility.

Which solution did I use previously and why did I switch?

At my last company, I used Nagios XI.

We brought on Centreon because we didn't have a monitoring solution.

How was the initial setup?

The initial setup is very easy. On the website, if you want to create an account, you go to the website for the cloud. There, you have the possibility to run a test of the system, before you want to implement it.

The implementation process was the same as the last system that I used. This made it easier to use.

What about the implementation team?

My company implemented the system. Then, we implemented a hub with the services, and this took five to six months. It takes time to implement all the services of a company because of all the firewalls. 

Only six to 12 people can currently configure Centreon.

What was our ROI?

Centreon helps me detect where the problem is quickly. When we resolve a problem quickly, this lowers our overall costs.

What's my experience with pricing, setup cost, and licensing?

I recommend looking at the trial version.

Which other solutions did I evaluate?

Centreon is better than Nagios XI in regards to cost and support response times, when you have a problem. If you have a problem, it costs money to contact the Nagios XI support.

What other advice do I have?

The solution is very good. I would rate it as a nine (out of 10). 

Disclosure: IT Central Station contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
Tchidat Linda
Engineer of Telecommunication at Gold Telecom
Real User
Top 5Leaderboard
Improved design, easy to learn, and open-sourced

Pros and Cons

  • "The solution's design has recently changed and it is visually pleasing with more color, for example, there is blue, black, and white."
  • "When using this solution in enterprise monitoring, you are able to see that there are some issues with equipment that could be causing a problem. Sometimes you want to make a root command that you do not want to be executed automatically. What we have tried to do is open an SSH session directly from the solution's interface but it is not possible."

What is our primary use case?

I use the solution for network monitoring. When I was in engineering school I used this solution and I found it really good. I wanted to bring the functionality of this solution to my community to show the engineers in my enterprise. There is not a lot of understanding of network monitoring and I am hoping to bring to them my knowledge. 

What is most valuable?

The solution's design has recently changed and it is visually pleasing with more color, for example, there is blue, black, and white. If you are a new user it is very simple, you do not need to be highly technical to use it. You can rapidly learn how to use this solution.

What needs improvement?

When using this solution in enterprise monitoring, you are able to see that there are some issues with equipment that could be causing a problem. Sometimes you want to make a root command that you do not want to be executed automatically. What we have tried to do is open an SSH session directly from the solution's interface but it is not possible. I have heard from others that there are scripts that can manage this task easily but I have not figured it out. In the area of the solution where you write the command, it is not very user-friendly and could be improved. We want the ability to write commands and execute them manually.

Additionally, I would like the ability to be able to have network mapping for this solution. I use network mapping to monitor networks and there are not very many tools that are able to do it or if there is a tool, it is not easy.

The color scheme of the interface could improve by allowing more variations or user customization.

For how long have I used the solution?

I have been using the solution for four years.

What do I think about the stability of the solution?

The solution is stable. 

What do I think about the scalability of the solution?

We have approximately 10 people using the solution in my company.

Which solution did I use previously and why did I switch?

In my community, we were previously using Nagios and Cacti.

What's my experience with pricing, setup cost, and licensing?

The solution is free. However, many open-sourced tools start out free but eventually start charging, I hope this solution stays free. 

What other advice do I have?

I rate Zabbix an eight out of ten.

Disclosure: I am a real user, and this review is based on my own experience and opinions.
Flag as inappropriate
Get our free report covering Zabbix, Paessler AG, Centreon, and other competitors of Nagios XI. Updated: November 2021.
554,382 professionals have used our research since 2012.