Kibana and Apache Druid Integration
Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.
5B+
Telegraf downloads
#1
Time series database
Source: DB Engines
1B+
Downloads of InfluxDB
2,800+
Contributors
Table of Contents
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Input and output integration overview
The Kibana plugin enables users to obtain status metrics from Kibana, a data visualization tool for Elasticsearch. By connecting to the Kibana API, this plugin captures various performance indicators and the health status of the Kibana service.
This plugin allows Telegraf to send JSON-formatted metrics to Apache Druid over HTTP, enabling real-time ingestion for analytical queries on high-volume time-series data.
Integration details
Kibana
The Kibana input plugin is designed to query the Kibana API to gather service status information. This plugin allows users to monitor their Kibana instances effectively by pulling metrics related to its health, performance, and operational metrics. By querying the Kibana API, this plugin provides insights into key parameters such as the current health status (green, yellow, red), uptime, heap memory usage, and request performance metrics. This information is crucial for administrators and operational teams looking to maintain optimal system performance and quickly address any issues that may arise. The configuration settings allow for flexible integration with other components in a microservices architecture, facilitating comprehensive monitoring solutions aligned with organizational needs, making it an essential tool for those leveraging the Elastic Stack in their infrastructure.
Apache Druid
This configuration uses Telegraf’s HTTP output plugin with json
data format to send metrics directly to Apache Druid, a real-time analytics database designed for fast, ad hoc queries on high-ingest time-series data. Druid supports ingestion via HTTP POST to various components like the Tranquility service or native ingestion endpoints. The JSON format is ideal for structuring Telegraf metrics into event-style records for Druid’s columnar and time-partitioned storage engine. Druid excels at powering interactive dashboards and exploratory queries across massive datasets, making it an excellent choice for real-time observability and monitoring analytics when integrated with Telegraf.
Configuration
Kibana
[[inputs.kibana]]
## Specify a list of one or more Kibana servers
servers = ["http://localhost:5601"]
## Timeout for HTTP requests
timeout = "5s"
## HTTP Basic Auth credentials
# username = "username"
# password = "pa$$word"
## Optional TLS Config
# tls_ca = "/etc/telegraf/ca.pem"
# tls_cert = "/etc/telegraf/cert.pem"
# tls_key = "/etc/telegraf/key.pem"
## Use TLS but skip chain & host verification
# insecure_skip_verify = false
## If 'use_system_proxy' is set to true, Telegraf will check env vars such as
## HTTP_PROXY, HTTPS_PROXY, and NO_PROXY (or their lowercase counterparts).
## If 'use_system_proxy' is set to false (default) and 'http_proxy_url' is
## provided, Telegraf will use the specified URL as HTTP proxy.
# use_system_proxy = false
# http_proxy_url = "http://localhost:8888"
Apache Druid
[[outputs.http]]
## Druid ingestion endpoint (e.g., Tranquility, HTTP Ingest, or Kafka REST Proxy)
url = "http://druid-ingest.example.com/v1/post"
## Use POST method to send events
method = "POST"
## Data format for Druid ingestion (expects JSON format)
data_format = "json"
## Optional headers (may vary depending on Druid setup)
# [outputs.http.headers]
# Content-Type = "application/json"
# Authorization = "Bearer YOUR_API_TOKEN"
## Optional timeout and TLS settings
timeout = "10s"
# tls_ca = "/path/to/ca.pem"
# tls_cert = "/path/to/cert.pem"
# tls_key = "/path/to/key.pem"
# insecure_skip_verify = false
Input and output integration examples
Kibana
-
Kibana Health Monitoring: Implement a dedicated dashboard to periodically poll the metrics from Kibana. This setup allows operations teams to have a real-time view of their Kibana instances’ health and metrics, enabling proactive performance management and immediate response capabilities in case of service degradation or failure.
-
Automated Alerting System: Integrate the metrics gathered from the Kibana plugin with an alerting system using tools like Prometheus or PagerDuty. By setting thresholds for key metrics (e.g., response time or heap usage), this integration can automatically notify the relevant personnel of performance issues, thereby reducing downtime and improving the response time for operational issues.
-
Resource Optimization Strategy: Use the memory usage and response time metrics collected by this plugin to formulate strategies for optimizing resource allocation in Kubernetes or other orchestration platforms. By analyzing trends over time, teams can adjust resource limits and requests dynamically, ensuring that Kibana instances function efficiently without over-provisioning resources.
Apache Druid
-
Real-Time Application Monitoring Dashboard: Use Telegraf to collect metrics from application servers and send them to Druid for immediate analysis and visualization in dashboards. Druid’s low-latency querying allows users to interactively explore system behavior in near real-time.
-
Security Event Aggregation: Aggregate and forward security-related metrics such as failed logins, port scans, or process anomalies to Druid. Analysts can build dashboards to monitor threat patterns and investigate incidents with millisecond-level granularity.
-
IoT Device Analytics: Collect telemetry from edge devices via Telegraf and send it to Druid for fast, scalable processing. Druid’s time-partitioned storage and roll-up capabilities are ideal for handling billions of small JSON events from sensors or gateways.
-
Web Traffic Behavior Exploration: Use Telegraf to capture web server metrics (e.g., requests per second, latency, error rates) and forward them to Druid. This enables teams to drill down into user behavior by region, device, or request type with subsecond query performance.
Feedback
Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Related Integrations
Related Integrations
HTTP and InfluxDB Integration
The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.
View IntegrationKafka and InfluxDB Integration
This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.
View IntegrationKinesis and InfluxDB Integration
The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.
View Integration