Docs » Configure application receivers » Configure application receivers for monitoring » Prometheus Velero

Prometheus Velero 🔗

Description 🔗

The Splunk Distribution of OpenTelemetry Collector provides this integration as the prometheus/velero monitor via the Smart Agent Receiver. This monitor gets metrics from Velero, which is an open source tool used to backup and restore, perform disaster recovery, and migrate Kubernetes cluster resources and persistent volumes. This monitor is a wrapper around the prometheus-exporter monitor that provides a restricted but expandable set of metrics.

Installation 🔗

This monitor is available in the SignalFx Smart Agent Receiver, which is part of the Splunk Distribution of OpenTelemetry Collector.

To install this integration:

  1. Deploy the Splunk Distribution of OpenTelemetry Collector to your host or container platform.

  2. Configure the monitor, as described in the next section.

Configuration 🔗

This Splunk Distribution of OpenTelemetry Collector allows embedding a Smart Agent monitor configuration in an associated Smart Agent Receiver instance.

Note: Providing a Prometheus Velero monitor entry in your Smart Agent or Collector configuration is required for its use. Use the appropriate form for your agent type.

Velero 🔗

Use the Helm chart to automatically enable Prometheus metrics for Velero.

Smart Agent 🔗

Configure service discovery on the Smart Agent, as shown in the following example:

monitors:
- type: prometheus/velero
  discoveryRule: container_image =~ "velero" && port == 8085
  port: 8085

To activate this monitor in the Smart Agent, add the following to your agent configuration:

monitors:  # All monitor config goes under this key
 - type: prometheus/velero
   ...  # Additional config

See Smart Agent example configuration for an autogenerated example of a YAML configuration file, with default values where applicable.

Splunk Distribution of OpenTelemetry Collector 🔗

To activate this monitor in the Splunk Distribution of OpenTelemetry Collector, add the following to your agent configuration:

receivers:
  smartagent/velero:
    type: prometheus/velero
    ...  # Additional config

See configuration examples for specific use cases that show how the collector can integrate and complement existing environments.

Configuration settings 🔗

The following table shows the configuration options for this monitor:

Option Required Type Description
httpTimeout no int64 HTTP timeout duration for both read and writes. Use a duration string that is accepted by https://golang.org/pkg/time/#ParseDuration (default: 10s)
username no string Basic Auth username to use on each request, if any.
password no string Basic Auth password to use on each request, if any.
useHTTPS no bool If true, the agent connects to the server using HTTPS instead of plain HTTP. (default: false)
httpHeaders no map of strings A map of HTTP header names to values. Comma separated multiple values for the same message-header is supported.
skipVerify no bool If useHTTPS is true and this option is also true, the exporter's TLS cert is not verified. (default: false)
caCertPath no string Path to the CA cert that has signed the TLS cert, unnecessary if skipVerify is set to false.
clientCertPath no string Path to the client TLS cert to use for TLS required connections
clientKeyPath no string Path to the client TLS key to use for TLS required connections
host yes string Host of the exporter
port yes integer Port of the exporter
useServiceAccount no bool Use pod service account to authenticate. (default: false)
metricPath no string Path to the metrics endpoint on the exporter server, usually /metrics (the default). (default: /metrics)
sendAllMetrics no bool Send all the metrics that come out of the Prometheus exporter without any filtering. This option has no effect when using the Prometheus exporter monitor directly since there is no built-in filtering, only when embedding it in other monitors. (default: false)

Metrics 🔗

These are the metrics available for this monitor.