etcd server 🔗
Description 🔗
The Splunk Distribution of OpenTelemetry Collector provides this integration as the etcd
monitor using the Smart Agent receiver.
This monitor reports etcd server metrics under the /metrics
path on its client port and optionally on locations given by --listen-metrics-urls
.
Installation 🔗
This monitor is available in the Smart Agent Receiver, which is part of the Splunk Distribution of OpenTelemetry Collector.
To install this integration:
Deploy the Splunk Distribution of OpenTelemetry Collector to your host or container platform.
Configure the monitor, as described in the next section.
Configuration 🔗
The Splunk Distribution of OpenTelemetry Collector allows embedding a Smart Agent monitor configuration in an associated Smart Agent Receiver instance.
Note: Providing an etcd monitor entry in your Collector or Smart Agent (deprecated) configuration is required for its use. Use the appropriate form for your agent type.
Splunk Distribution of OpenTelemetry Collector 🔗
To activate this monitor in the OpenTelemetry Collector, add the following to your agent configuration:
receivers:
smartagent/etcd:
type: etcd
... # Additional config
To complete the monitor activation, you must also include the smartagent/etcd
receiver item in a metrics
pipeline. To do this, add the receiver item to the service
> pipelines
> metrics
> receivers
section of your configuration file.
See configuration examples for specific use cases that show how the Splunk Distribution of OpenTelemetry Collector can integrate and complement existing environments.
Smart Agent 🔗
To activate this monitor in the Smart Agent, add the following to your agent configuration. Note that this configuration assumes that the client certificate and key are accessible by the Smart Agent in the specified path:
monitors: # All monitor config goes under this key
- type: etcd
... # Additional config
See Smart Agent example configuration for an autogenerated example of a YAML configuration file, with default values where applicable.
Example configurations 🔗
The following is an example configuration for this monitor:
monitors:
- type: etcd
discoveryRule: kubernetes_pod_name =~ "etcd" && target == "pod"
port: 2379
useHTTPS: true
skipVerify: true
sendAllMetrics: true
clientCertPath: /var/lib/minikube/certs/etcd/server.crt
clientKeyPath: /var/lib/minikube/certs/etcd/server.key
extraDimensions:
metric_source: etcd
The following table shows the configuration options for this monitor:
Option |
Required |
Type |
Description |
---|---|---|---|
|
no |
|
HTTP timeout duration for both read and writes. This should be a duration string that is accepted by https://golang.org/pkg/time/#ParseDuration (default: |
|
no |
|
Basic Auth username to use on each request, if any. |
|
no |
|
Basic Auth password to use on each request, if any. |
|
no |
|
If |
|
no |
|
A map of HTTP header names to values. Comma separated multiple values for the same message-header is supported. |
|
no |
|
If useHTTPS is |
|
no |
|
Path to the CA cert that has signed the TLS cert, unnecessary if |
|
no |
|
Path to the client TLS cert to use for TLS required connections |
|
no |
|
Path to the client TLS key to use for TLS required connections |
|
yes |
|
Host of the exporter |
|
yes |
|
Port of the exporter |
|
no |
|
Use pod service account to authenticate. (default: |
|
no |
|
Path to the metrics endpoint on the exporter server, usually |
|
no |
|
Send all the metrics that come out of the Prometheus exporter without any filtering. This option has no effect when using the prometheus exporter monitor directly since there is no built-in filtering, only when embedding it in other monitors. (default: |
Metrics 🔗
The following metrics are available for this integration:
Get help 🔗
If you are not able to see your data in Splunk Observability Cloud, try these tips:
Submit a case in the Splunk Support Portal
Available to Splunk Observability Cloud customers
-
Available to Splunk Observability Cloud customers
Ask a question and get answers through community support at Splunk Answers
Available to Splunk Observability Cloud customers and free trial users
Join the Splunk #observability user group Slack channel to communicate with customers, partners, and Splunk employees worldwide
Available to Splunk Observability Cloud customers and free trial users
To learn how to join, see Get Started with Splunk Community - Chat groups
To learn about even more support options, see Splunk Customer Success.