Cisco Model-Driven Telemetry and Prometheus Integration

Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.

info

This is not the recommended configuration for real-time query at scale. For query and compression optimization, high-speed ingest, and high availability, you may want to consider Cisco MDT and InfluxDB.

5B+

Telegraf downloads

#1

Time series database
Source: DB Engines

1B+

Downloads of InfluxDB

2,800+

Contributors

Table of Contents

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Input and output integration overview

The Cisco Model-Driven Telemetry (MDT) plugin facilitates the collection of telemetry data from Cisco networking platforms, utilizing gRPC and TCP transport mechanisms. This plugin is essential for users looking to implement advanced telemetry solutions for better insights and operational efficiency.

The Prometheus Output Plugin enables Telegraf to expose metrics at an HTTP endpoint for scraping by a Prometheus server. This integration allows users to collect and aggregate metrics from various sources in a format that Prometheus can process efficiently.

Integration details

Cisco Model-Driven Telemetry

Cisco model-driven telemetry (MDT) is designed to provide a robust means of consuming telemetry data from various Cisco platforms, including IOS XR, IOS XE, and NX-OS. This plugin focuses on the efficient transport of telemetry data using either TCP or gRPC protocols, offering flexibility based on the network environment and requirements. The gRPC transport is particularly advantageous as it supports TLS for enhanced security through encryption and authentication. The plugin is compatible with a range of software versions on Cisco devices, enabling organizations to leverage telemetry capabilities across their network operations. It is especially useful for network monitoring and analytics, as it enables real-time data collection directly from Cisco devices, enhancing visibility into network performance, resource utilization, and operational metrics.

Prometheus

This plugin for facilitates the integration with Prometheus, a well-known open-source monitoring and alerting toolkit designed for reliability and efficiency in large-scale environments. By working as a Prometheus client, it allows users to expose a defined set of metrics via an HTTP server that Prometheus can scrape at specified intervals. This plugin plays a crucial role in monitoring diverse systems by allowing them to publish performance metrics in a standardized format, enabling extensive visibility into system health and behavior. Key features include support for configuring various endpoints, enabling TLS for secure communication, and options for HTTP basic authentication. The plugin also integrates seamlessly with global Telegraf configuration settings, supporting extensive customization to fit specific monitoring needs. This promotes interoperability in environments where different systems must communicate performance data effectively. Leveraging Prometheus’s metric format, it allows for flexible metric management through advanced configurations such as metric expiration and collectors control, offering a sophisticated solution for monitoring and alerting workflows.

Configuration

Cisco Model-Driven Telemetry

[[inputs.cisco_telemetry_mdt]]
 ## Telemetry transport can be "tcp" or "grpc".  TLS is only supported when
 ## using the grpc transport.
 transport = "grpc"

 ## Address and port to host telemetry listener
 service_address = ":57000"

 ## Grpc Maximum Message Size, default is 4MB, increase the size. This is
 ## stored as a uint32, and limited to 4294967295.
 max_msg_size = 4000000

 ## Enable TLS; grpc transport only.
 # tls_cert = "/etc/telegraf/cert.pem"
 # tls_key = "/etc/telegraf/key.pem"

 ## Enable TLS client authentication and define allowed CA certificates; grpc
 ##  transport only.
 # tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]

 ## Define (for certain nested telemetry measurements with embedded tags) which fields are tags
 # embedded_tags = ["Cisco-IOS-XR-qos-ma-oper:qos/interface-table/interface/input/service-policy-names/service-policy-instance/statistics/class-stats/class-name"]

 ## Include the delete field in every telemetry message.
 # include_delete_field = false

 ## Specify custom name for incoming MDT source field.
 # source_field_name = "mdt_source"

 ## Define aliases to map telemetry encoding paths to simple measurement names
 [inputs.cisco_telemetry_mdt.aliases]
   ifstats = "ietf-interfaces:interfaces-state/interface/statistics"
 ## Define Property Xformation, please refer README and https://pubhub.devnetcloud.com/media/dme-docs-9-3-3/docs/appendix/ for Model details.
 [inputs.cisco_telemetry_mdt.dmes]
#    Global Property Xformation.
#    prop1 = "uint64 to int"
#    prop2 = "uint64 to string"
#    prop3 = "string to uint64"
#    prop4 = "string to int64"
#    prop5 = "string to float64"
#    auto-prop-xfrom = "auto-float-xfrom" #Xform any property which is string, and has float number to type float64
#    Per Path property xformation, Name is telemetry configuration under sensor-group, path configuration "WORD         Distinguished Name"
#    Per Path configuration is better as it avoid property collision issue of types.
#    dnpath = '{"Name": "show ip route summary","prop": [{"Key": "routes","Value": "string"}, {"Key": "best-paths","Value": "string"}]}'
#    dnpath2 = '{"Name": "show processes cpu","prop": [{"Key": "kernel_percent","Value": "float"}, {"Key": "idle_percent","Value": "float"}, {"Key": "process","Value": "string"}, {"Key": "user_percent","Value": "float"}, {"Key": "onesec","Value": "float"}]}'
#    dnpath3 = '{"Name": "show processes memory physical","prop": [{"Key": "processname","Value": "string"}]}'

 ## Additional GRPC connection settings.
 [inputs.cisco_telemetry_mdt.grpc_enforcement_policy]
  ## GRPC permit keepalives without calls, set to true if your clients are
  ## sending pings without calls in-flight. This can sometimes happen on IOS-XE
  ## devices where the GRPC connection is left open but subscriptions have been
  ## removed, and adding subsequent subscriptions does not keep a stable session.
  # permit_keepalive_without_calls = false

  ## GRPC minimum timeout between successive pings, decreasing this value may
  ## help if this plugin is closing connections with ENHANCE_YOUR_CALM (too_many_pings).
  # keepalive_minimum_time = "5m"

Prometheus

[[outputs.prometheus_client]]
  ## Address to listen on.
  ##   ex:
  ##     listen = ":9273"
  ##     listen = "vsock://:9273"
  listen = ":9273"

  ## Maximum duration before timing out read of the request
  # read_timeout = "10s"
  ## Maximum duration before timing out write of the response
  # write_timeout = "10s"

  ## Metric version controls the mapping from Prometheus metrics into Telegraf metrics.
  ## See "Metric Format Configuration" in plugins/inputs/prometheus/README.md for details.
  ## Valid options: 1, 2
  # metric_version = 1

  ## Use HTTP Basic Authentication.
  # basic_username = "Foo"
  # basic_password = "Bar"

  ## If set, the IP Ranges which are allowed to access metrics.
  ##   ex: ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
  # ip_range = []

  ## Path to publish the metrics on.
  # path = "/metrics"

  ## Expiration interval for each metric. 0 == no expiration
  # expiration_interval = "60s"

  ## Collectors to enable, valid entries are "gocollector" and "process".
  ## If unset, both are enabled.
  # collectors_exclude = ["gocollector", "process"]

  ## Send string metrics as Prometheus labels.
  ## Unless set to false all string metrics will be sent as labels.
  # string_as_label = true

  ## If set, enable TLS with the given certificate.
  # tls_cert = "/etc/ssl/telegraf.crt"
  # tls_key = "/etc/ssl/telegraf.key"

  ## Set one or more allowed client CA certificate file names to
  ## enable mutually authenticated TLS connections
  # tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]

  ## Export metric collection time.
  # export_timestamp = false

  ## Specify the metric type explicitly.
  ## This overrides the metric-type of the Telegraf metric. Globbing is allowed.
  # [outputs.prometheus_client.metric_types]
  #   counter = []
  #   gauge = []

Input and output integration examples

Cisco Model-Driven Telemetry

  1. Real-Time Network Monitoring: Utilize the Cisco MDT plugin to collect network performance metrics from Cisco routers and switches. By feeding telemetry data into a visualization tool, network operators can observe traffic trends, bandwidth usage, and error rates in real-time. This proactive monitoring allows teams to swiftly address issues before they affect network performance, resulting in a more reliable service.

  2. Automated Anomaly Detection: Integrate Cisco MDT with machine learning algorithms to create an automated anomaly detection system. By continuously analyzing telemetry data, the system can identify deviations from typical operational patterns, providing alerts for unusual conditions that may signify network problems or security threats, which can aid in maintaining operational integrity.

  3. Dynamic Configuration Management: Leveraging the telemetry data collected from Cisco devices, organizations can implement dynamic configuration management solutions that automatically adjust network settings based on current performance indicators. For instance, if the telemetry indicates high utilization on certain links, the system could dynamically route traffic to underutilized paths, optimizing resource usage.

  4. Enhanced Reporting and Analytics: Use the Cisco MDT plugin to feed detailed telemetry data into analytics platforms, enabling comprehensive reporting on network health and performance. Historical and real-time analysis can guide decision-making and strategic planning, helping organizations to allocate resources more effectively and understand their network’s operational landscape better.

Prometheus

  1. Monitoring Multi-cloud Deployments: Utilize the Prometheus plugin to collect metrics from applications running across multiple cloud providers. This scenario allows teams to centralize monitoring through a single Prometheus instance that scrapes metrics from different environments, providing a unified view of performance metrics across hybrid infrastructures. It streamlines reporting and alerting, enhancing operational efficiency without needing complex integrations.

  2. Enhancing Microservices Visibility: Implement the plugin to expose metrics from various microservices within a Kubernetes cluster. Using Prometheus, teams can visualize service metrics in real time, identify bottlenecks, and maintain system health checks. This setup supports adaptive scaling and resource utilization optimization based on insights generated from the collected metrics. It enhances the ability to troubleshoot service interactions, significantly improving the resilience of the microservice architecture.

  3. Real-time Anomaly Detection in E-commerce: By leveraging this plugin alongside Prometheus, an e-commerce platform can monitor key performance indicators such as response times and error rates. Integrating anomaly detection algorithms with scraped metrics allows the identification of unexpected patterns indicating potential issues, such as sudden traffic spikes or backend service failure. This proactive monitoring empowers business continuity and operational efficiency, minimizing potential downtimes while ensuring service reliability.

  4. Performance Metrics Reporting for APIs: Utilize the Prometheus Output Plugin to gather and report API performance metrics, which can then be visualized in Grafana dashboards. This use case enables detailed analysis of API response times, throughput, and error rates, promoting continuous improvement of API services. By closely monitoring these metrics, teams can quickly react to degradation, ensuring optimal API performance and maintaining a high level of service availability.

Feedback

Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Related Integrations

HTTP and InfluxDB Integration

The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.

View Integration

Kafka and InfluxDB Integration

This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.

View Integration

Kinesis and InfluxDB Integration

The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.

View Integration