Grafana

Grafana

DevOps / Monitoring / Monitoring Tools
Needs advice
on
DatadogDatadog
and
GrafanaGrafana

Hello :) We are using Datadog on Kong to monitor the metrics and analytics.

We feel that the cost associated with Datadog is high in terms of custom metrics and indexations. So, we planned to find an alternative for Datadog and we are looking into Grafana implementation with kong.

Will the shift from Datadog to Grafana be a wise move and flawless?

READ MORE
5 upvotes·61K views
Replies (1)
Software Engineer ·

Hi Sathish,

Straightforward responding you: Will the shift from Datadog to Grafana be a wise move? R: Maybe, specially in the long run.

and flawless? R: It depends, but I don't think so. And it also won't be so smooth. You'll reach a point of Datatog deprec where your team will be supporting both, and at this peak will be when you'll think it was a bad move, because you'll be paying for Datadog and Prometheus/Grafana servers, plus your team will have to work on new system metrics and fine tuning metric queries and graphs on Grafana, so you can expect a delay on business deliveries. Your team will feel overburdened.

So bear in mind that basically you'll be moving from an out of the box monitoring solution to your own. The analogy I'd do is to moving out from a cloud provider to your own self managed servers, though way less complex of course.

Grafana itself won't do the trick. You'll need a data scrapper system like Prometheus to collect metrics/data from your app, or use a middleware Pushgateway to receive these data so Prometheus can scrape them. To do this you'll need some backend work to expose metrics data, instead of HTTP post your metrics data to Datadog, so you can expect a little re-architecture or engineering on that, depending on how your system is designed.

In the long run, you'll probably need a team focused only on that: to take care of the Prometheus/Grafana updates as well as their servers (after all they're still systems that need to be managed).

I think all of this will depend on the size of your company now and your teams (which I advise you to include them on that decision), and your budget plus rush for business deliveries. In the long run it'll pay, for your company will be more specialized on this stack, and you can have a team focused on that plus developer support.

If you wanna go with it, I'd suggest to do a small PoC (which still will need a good amount of work). Select a project (or even an endpoint/ mobile feature) to have both integrations: Datadog plus Grafana, and do the same graphs you have on Datadog to Grafana. I think this way you can have a better picture of the reality on how that will be for your company.

READ MORE
1 upvote·1 comment·1.2K views
Giorgio Torres
Giorgio Torres
·
June 9th 2023 at 9:46AM

I also posted your question on ChatGPT, and here's its full response on that, which I also agree:

```Shifting from Datadog to Grafana for monitoring metrics and analytics can be a reasonable alternative, depending on your specific requirements and use case. Grafana is a popular open-source platform that provides rich visualization and monitoring capabilities. However, it's important to note that Grafana and Datadog have different feature sets and focus areas. Here are a few considerations to help you evaluate the potential shift:

1. Feature Comparison: Evaluate the features and functionalities offered by Grafana and compare them to the ones you currently use in Datadog. Ensure that Grafana meets your specific monitoring, analytics, and visualization needs.

2. Custom Metrics and Indexing: Understand your requirements for custom metrics and indexing in Grafana. Ensure that Grafana provides the necessary flexibility and support for the metrics and indexes you require. Check if Grafana integrates well with Kong and offers plugins or extensions for capturing custom metrics.

3. Learning Curve: Consider the learning curve associated with Grafana, as your team will need to become familiar with its interface, configuration, and query language (e.g., Prometheus Query Language). Assess whether your team has the required expertise or if additional training will be necessary.

4. Data Storage and Scalability: Evaluate how Grafana handles data storage and scalability. Determine if it can handle the volume of metrics and analytics data generated by Kong and if it integrates seamlessly with your existing data storage solutions.

5. Alerting and Notification: Determine if Grafana's alerting and notification features meet your requirements. Check if it supports real-time alerting, customizable thresholds, and various notification channels (e.g., email, Slack).

6. Ecosystem and Community Support: Consider the ecosystem and community support around Grafana. It has a large user community, which means you can find resources, plugins, and community-contributed dashboards to enhance your monitoring capabilities.

7. Cost Comparison: Evaluate the cost aspect of Grafana compared to Datadog. While Grafana itself is open source, you need to consider the costs associated with additional components, such as data storage, alerting systems, and any necessary infrastructure changes. Calculate the total cost of ownership and compare it to your current expenses with Datadog.

It's important to conduct a thorough evaluation and possibly run a proof of concept or pilot project to ensure that Grafana meets your specific needs before fully transitioning from Datadog.```

·
Reply

Hello People, I want suggestions about monitoring and alert tools to use with .NET application which will be hosted on Microsoft Azure. I have Azure Monitor,, Grafana, and Prometheus in my consideration. What would you suggest among these tools? If you have any other suggestions, please share the ideas. Thank you.

READ MORE
7 upvotes·31.5K views
Replies (1)
Needs advice
on
Azure MonitorAzure Monitor
and
PrometheusPrometheus

Can I get metrics available through Prometheus into Azure Monitor, specifically into log analytics? (VM'S). I am running a couple of VM's inside Azure portal and I have my own private besu nodes running on them. I have my metrics set up inside the Prometheus but I was hoping to hook it up securely to Grafana but I tried everything and I can't. So the next thing is to see if can I get the metrics available through Prometheus into azure monitor, specifically into log analytics. The aim is to get the sync status, and the highest block number on each node, into log analytics so we can see what each is doing. That way we know, on a quick look, the status of each node and by extension, the condition of the private chain. What worries me is that although I have alerts if blocks stop being created or nodes lose peers we cannot see it quickly.

Prometheus is one option to give us those stats. If we can get data from Prometheus into log analytics that would solve the problem.

Can anyone help me with how I can go about it or any links? All I am seeing is for containers but I want for my VMs.

READ MORE
4 upvotes·48.5K views
Replies (2)
Systems Engineer ·

Hi Darragh,

You can use Function Apps to scrape metrics and send them to a Log Analytics Data Collector API. Still, from my experience, I can say that it's hard to query data in Prometheus format with KQL. Instead, I'd add the Application Insights instrumentation to the application or proceed with building the Prometheus setup.

READ MORE
2 upvotes·3.8K views

Hi Darragh, I'm part of the User Success team at Grafana. I'd love to hear more about what you're working on, understand your use case and see if I can provide more resources for you. Care to chat?

READ MORE
1 upvote·4K views
Needs advice
on
GolangGolangGrafanaGrafana
and
LogstashLogstash

Hi everyone. I'm trying to create my personal syslog monitoring.

  1. To get the logs, I have uncertainty to choose the way: 1.1 Use Logstash like a TCP server. 1.2 Implement a Go TCP server.

  2. To store and plot data. 2.1 Use Elasticsearch tools. 2.2 Use InfluxDB and Grafana.

I would like to know... Which is a cheaper and scalable solution?

Or even if there is a better way to do it.

READ MORE
10 upvotes·200.2K views
Replies (3)
Recommends
on
Grafana
Loki

Hi Juan

A very simple and cheap (resource usage) option here would be to use promtail to send syslog data to Loki and visualise Loki with Grafana using the native Grafana Loki data source. I have recently put together this set up and promtail and Loki are less resource intensive than Logstash/ES and it is a simple set up and configuration and works very nicely.

READ MORE
4 upvotes·2 comments·4.5K views
Sunil Chaudhari
Sunil Chaudhari
·
October 27th 2021 at 1:23AM

Hi,

Does promtel available for PCF?

·
Reply
Gary Wilson
Gary Wilson
·
October 27th 2021 at 1:38PM

Hi @sunilmchaudhari I do not know. I assume by PCF you are refering to Pivot Cloud Foundry, which I have no knowledge of sorry. Promtail is a go binary so if you can add log data to a syslog, then you can process it with Promtail.

·
Reply
Principal Software Architect at Breu Inc.·
Recommends
on
M3
Prometheus

Take a look at Prometheus or m3 as a storage engine.

READ MORE
3 upvotes·3.2K views
View all (3)
Needs advice
on
PrometheusPrometheus
and
ThanosThanos

Hi All, We have Thanos sidecar and Prometheus set up in GCP and a Prometheus server in AWS. we want to push all the metrics of GCP to AWS. We will be creating a VPN link-up between them. But then how the GCP metrics would be connected with Prometheus/ Grafana.? I mean which IP to use it for this.?

READ MORE
3 upvotes·43.5K views
Replies (1)
Recommends
on
Prometheus
Thanos

Hi Suraj,

If :

  1. Thanos is installed on AWS
  2. The Prometheus on GCP is configured with a Thanos sidecar forwarding metrics to the AWS Thanos.
  3. The Prometheus on AWS is configured with a Thanos sidecar forwarding metrics to the AWS Thanos.

Then :

Your Grafana should then point to the Thanos Querier IP:port on AWS. You will then be able to view the GCP metrics as well as the AWS metrics.

Trust this helps. Send me an email if you need more help in this area.

Scott Fulton scott.fulton@opscruise.com

READ MORE
5 upvotes·459 views
Senior Software Engineering Manager at PayIt·

Grafana and Prometheus together, running on Kubernetes , is a powerful combination. These tools are cloud-native and offer a large community and easy integrations. At PayIt we're using exporting Java application metrics using a Dropwizard metrics exporter, and our Node.js services now use the prom-client npm library to serve metrics.

READ MORE
With multiplying microservices running on Kubernetes, PayIt turned to Grafana and Prometheus for observability at cloud native scale | Grafana Labs (grafana.com)
16 upvotes·1M views
Senior SRE at African Bank·
Needs advice
on
GrafanaGrafanaGraphiteGraphite
and
PrometheusPrometheus

Looking for a tool which can be used for mainly dashboard purposes, but here are the main requirements:

  • Must be able to get custom data from AS400,
  • Able to display automation test results,
  • System monitoring / Nginx API,
  • Able to get data from 3rd parties DB.

Grafana is almost solving all the problems, except AS400 and no database to get automation test results.

READ MORE
4 upvotes·831.1K views
Replies (1)
Technical Specialist, Software Engineering at AT&T·
Recommends
on
Grafana
Prometheus

You can look out for Prometheus Instrumentation (https://prometheus.io/docs/practices/instrumentation/) Client Library available in various languages https://prometheus.io/docs/instrumenting/clientlibs/ to create the custom metric you need for AS4000 and then Grafana can query the newly instrumented metric to show on the dashboard.

READ MORE
3 upvotes·1 comment·616.6K views
Susmita Meher
Susmita Meher
·
October 3rd 2020 at 8:01PM

Thank you for the suggestions.

However, I managed to write libraries for Prometheus using NodeJs for the adhoc quesries.

·
Reply
Needs advice
on
DatadogDatadogNew RelicNew Relic
and
SysdigSysdig

We are looking for a centralised monitoring solution for our application deployed on Amazon EKS. We would like to monitor using metrics from Kubernetes, AWS services (NeptuneDB, AWS Elastic Load Balancing (ELB), Amazon EBS, Amazon S3, etc) and application microservice's custom metrics.

We are expected to use around 80 microservices (not replicas). I think a total of 200-250 microservices will be there in the system with 10-12 slave nodes.

We tried Prometheus but it looks like maintenance is a big issue. We need to manage scaling, maintaining the storage, and dealing with multiple exporters and Grafana. I felt this itself needs few dedicated resources (at least 2-3 people) to manage. Not sure if I am thinking in the correct direction. Please confirm.

You mentioned Datadog and Sysdig charges per host. Does it charge per slave node?

READ MORE
7 upvotes·1.5M views
Replies (3)
Recommends
on
Datadog

Can't say anything to Sysdig. I clearly prefer Datadog as

  • they provide plenty of easy to "switch-on" plugins for various technologies (incl. most of AWS)
  • easy to code (python) agent plugins / api for own metrics
  • brillant dashboarding / alarms with many customization options
  • pricing is OK, there are cheaper options for specific use cases but if you want superior dashboarding / alarms I haven't seen a good competitor (despite your own Prometheus / Grafana / Kibana dog food)

IMHO NewRelic is "promising since years" ;) good ideas but bad integration between their products. Their Dashboard query language is really nice but lacks critical functions like multiple data sets or advanced calculations. Needless to say you get all of that with Datadog.

Need help setting up a monitoring / logging / alarm infrastructure? Send me a message!

READ MORE
10 upvotes·2 comments·416.3K views
Medeti Vamsi Krishna
Medeti Vamsi Krishna
·
June 30th 2020 at 11:52AM

Thanks for the reply, I am working on DataDog trail version now. I am able to see my containers/pods/VMs metrics in the DataDog.

I am trying to do the jmx integration with autodiscovery now. But I am not able to see the jvm metrics in DataDog. Can you please help on this?

Here is my deployment yaml:

`

apiVersion: apps/v1

kind: Deployment

metadata:

name: myapp

namespace: datadog

annotations:

ad.datadoghq.com/myapp.check_names: >-

'["myapp"]'

ad.datadoghq.com/myapp.init_configs: >-

'[{"is_jmx": true, "collect_default_metrics": true}]'

ad.datadoghq.com/tomcat.instances: >-

'[{"host": "%%host%%","port":"5000"}]'

labels:

app: myapp

spec:

selector:

matchLabels:

app: myapp

template:

metadata:

labels:

app: myapp

spec:

containers:

- name: myapp

image: nexus.nslhub.com/sample-java-app:2.0

imagePullPolicy: Always

ports:

- containerPort: 8080

name: http

- containerPort: 5000

name: jmx

imagePullSecrets:

- name: myappsecret

nodeSelector:

kubernetes.io/hostname: ip-10-5-7-173.ap-south-1.compute.internal

`

·
Reply
Jens Günther
Jens Günther
·
June 30th 2020 at 11:57AM

Would like to help, but there could be hundreds of reasons why the incoming and outgoing jmx ports are not accessible from the agent.

·
Reply
Recommends
on
Instana

Hi Medeti,

you are right. Building based on your stack something with open source is heavy lifting. A lot of people I know start with such a set-up, but quickly run into frustration as they need to dedicated their best people to build a monitoring which is doing the job in a professional way.

As you are microservice focussed and are looking for 'low implementation and maintenance effort', you might want to have a look at INSTANA, which was built with modern tool stacks in mind. https://www.instana.com/apm-for-microservices/

We have a public sand-box available if you just want to have a look at the product once and of course also a free-trial: https://www.instana.com/getting-started-with-apm/

Let me know if you need anything on top.

READ MORE
8 upvotes·416.2K views
View all (3)
Team Lead at XYZ·
Needs advice
on
MetricbeatMetricbeat
and
PrometheusPrometheus

Hi, We have a situation, where we are using Prometheus to get system metrics from PCF (Pivotal Cloud Foundry) platform. We send that as time-series data to Cortex via a Prometheus server and built a dashboard using Grafana. There is another pipeline where we need to read metrics from a Linux server using Metricbeat, CPU, memory, and Disk. That will be sent to Elasticsearch and Grafana will pull and show the data in a dashboard.

Is it OK to use Metricbeat for Linux server or can we use Prometheus?

What is the difference in system metrics sent by Metricbeat and Prometheus node exporters?

Regards, Sunil.

READ MORE
2 upvotes·569.7K views
Replies (2)
Recommends
on
Prometheus

If you're already using Prometheus for your system metrics, then it seems like standing up Elasticsearch just for Linux host monitoring is excessive. The node_exporter is probably sufficient if you'e looking for standard system metrics.

Another thing to consider is that Metricbeat / ELK use a push model for metrics delivery, whereas Prometheus pulls metrics from each node it is monitoring. Depending on how you manage your network security, opting for one solution over two may make things simpler.

READ MORE
5 upvotes·1 comment·349K views
Manish Sharma
Manish Sharma
·
July 23rd 2021 at 9:41AM

This is perfect answer.

·
Reply
CEO at Scrayos UG (haftungsbeschränkt)·

Grafana is used in combination with Prometheus to display the gathered stats and to monitor our physical servers aswell as their virtual applications. While Grafana also allows to configure automated alerts and rules, we decided to use Prometheus Alertmanager, as it is offers advanced features for silences (muting of alerts for a specific time) and also allows more fine-grained rules and notifications for each alert.

READ MORE
2 upvotes·69.6K views