Salesforce and SigNoz Integration

Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.

info

This is not the recommended configuration for real-time query at scale. For query and compression optimization, high-speed ingest, and high availability, you may want to consider Salesforce and InfluxDB.

5B+

Telegraf downloads

#1

Time series database
Source: DB Engines

1B+

Downloads of InfluxDB

2,800+

Contributors

Table of Contents

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Input and output integration overview

The Salesforce Telegraf plugin collects crucial metrics regarding the API usage and limits in Salesforce organizations, enabling effective monitoring and management of API consumption.

This configuration turns any Telegraf agent into a Remote Write publisher for SigNoz, streaming rich metrics straight into the SigNoz backend with a single URL change.

Integration details

Salesforce

The Salesforce plugin allows users to gather metrics about API usage limits and the remaining usage within their Salesforce organization. By leveraging Salesforce’s REST API, specifically the limits endpoint, this plugin provides critical insights into how much of the API usage has been consumed and what remains available. This is particularly important for organizations that rely on Salesforce for their operations, as exceeding API limits can interrupt service and hinder business processes. The plugin processes data into a structured format containing maximum and remaining values for various API operations, making it easier for teams to monitor their usage and plan accordingly. The provided configuration allows users to customize their credentials, environment type (sandbox or production), and API version, ensuring flexibility in different deployment scenarios.

SigNoz

SigNoz is an open source observability platform that stores metrics, traces, and logs. When you deploy SigNoz, its signoz-otel-collector-metrics service exposes a Prometheus Remote Write receiver (default :13133/api/v1/write). By configuring Telegraf’s Prometheus plugin to point at this endpoint, you can push any Telegraf collected metrics, SNMP counters, cloud services, or business KPIs—directly into SigNoz. The plugin natively serializes metrics in the Remote Write protobuf format, supports external labels, metadata export, retries, and TLS or bearer-token auth, so it fits zero-trust and multi-tenant SigNoz clusters. Inside SigNoz, the data lands in ClickHouse tables that back Metrics Explorer, alert rules, and unified dashboards. This approach lets organizations unify Prometheus and OTLP pipelines, enables long-term retention powered by ClickHouse compression, and avoids vendor lock-in while retaining PromQL-style queries.

Configuration

Salesforce

[[inputs.salesforce]]
  ## specify your credentials
  ##
  username = "your_username"
  password = "your_password"
  ##
  ## (optional) security token
  # security_token = "your_security_token"
  ##
  ## (optional) environment type (sandbox or production)
  ## default is: production
  ##
  # environment = "production"
  ##
  ## (optional) API version (default: "39.0")
  ##
  # version = "39.0"

SigNoz

[[outputs.prometheusremotewrite]]
  ## SigNoz OTEL-Collector metrics endpoint (Prometheus Remote Write receiver)
  ## Default port is 13133 when you install SigNoz with the Helm chart
  url = "http://signoz-otel-collector-metrics.monitoring.svc.cluster.local:13133/api/v1/write"

  ## Add identifying labels so you can slice & dice the data later
  external_labels = { host = "${HOSTNAME}", agent = "telegraf" }

  ## Forward host metadata for richer dashboards (SigNoz maps these to ClickHouse columns)
  send_metadata = true

  ## ----- Authentication (comment out what you don’t need) -----
  # bearer_token   = "$SIGNOZ_TOKEN"          # SaaS tenant token
  # basic_username = "signoz"                 # Basic auth (self-hosted)
  # basic_password = "secret"

  ## ----- TLS options (for SaaS or HTTPS self-hosted) -----
  # tls_ca                  = "/etc/ssl/certs/ca.crt"
  # tls_cert                = "/etc/telegraf/certs/telegraf.crt"
  # tls_key                 = "/etc/telegraf/certs/telegraf.key"
  # insecure_skip_verify    = false

  ## ----- Performance tuning -----
  max_batch_size = 10000      # samples per POST
  timeout        = "10s"
  retry_max      = 3

Input and output integration examples

Salesforce

  1. Monitoring API Limit Usage for Scaling Decisions: Use the Salesforce plugin to track API limit usage over time and make informed decisions about when to scale Salesforce resources. By visualizing API consumption patterns, organizations can predict peak usage times, allowing them to proactively adjust their infrastructure or request higher limits as needed. This optimization leads to better performance and less downtime during critical business operations.

  2. Automated Alert System for API Limit Exceedance: Integrate this plugin with a notification system to alert teams when API usage approaches critical limits. This setup not only ensures teams are proactively notified to prevent disruptions, but also helps in maintaining operational continuity and customer satisfaction. The alerts can be configured to trigger automated scripts that either adjust load or inform stakeholders accordingly.

  3. Comparative Analysis of Multiple Salesforces: Leverage the Salesforce Input Plugin to gather metrics from multiple Salesforce instances across different departments or business units. By centralizing this data, organizations can perform comparative analyses to identify departments that may be exceeding their API limits more frequently than others. This allows for targeted discussions and strategies to balance API usage across the organization, leading to better resource allocation and efficiency.

SigNoz

  1. Multi-Cluster Federated Monitoring: Drop a Telegraf DaemonSet into each Kubernetes cluster, tag metrics with cluster=<name>, and Remote Write them to a central SigNoz instance. Ops teams get a single PromQL window across prod, staging, and edge clusters without running Thanos sidecars.

  2. Factory-Floor Edge Gateway: A rugged Intel NUC on the shop floor runs Telegraf to scrape Modbus PLCs and environmental sensors. It batches readings every 5 seconds and pushes them over an intermittent 4G link to SigNoz SaaS. ClickHouse compression keeps costs low while AI-based outlier detection in SigNoz flags overheating motors before failure.

  3. SaaS Usage Metering: Telegraf runs alongside each micro-service, exporting per-tenant counters (api_calls, gigabytes_processed). Remote Write streams the data to SigNoz where a scheduled ClickHouse materialized view aggregates usage for monthly billing—no separate metering stack required.

  4. Autoscaling Feedback Loop: Combine Telegraf’s Kubernetes input with the Remote Write output to publish granular pod CPU and queue-length metrics into SigNoz. A custom SigNoz alert fires when P95 latency breaches 200 ms and a GitOps controller reads that alert to trigger a HorizontalPodAutoscaler tweak—closing the loop between observability and automation.

Feedback

Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Related Integrations

HTTP and InfluxDB Integration

The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.

View Integration

Kafka and InfluxDB Integration

This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.

View Integration

Kinesis and InfluxDB Integration

The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.

View Integration