Datadog Metrics Sink
The Vector datadog_metrics
sink
sends metrics to Datadog metrics.
Configuration
- Common
- Advanced
- vector.toml
- vector.yaml
- vector.json
[sinks.my_sink_id]# Generaltype = "datadog_metrics" # requiredinputs = ["my-source-or-transform-id", "prefix-*"] # requiredapi_key = "${DATADOG_API_KEY_ENV_VAR}" # requireddefault_namespace = "service" # optional, no default# Healthcheckhealthcheck.enabled = true # optional, default
- commonrequiredstring
api_key
Datadog API key
- Syntax:
literal
- View examples
- Syntax:
- optionaltable
batch
Configures the sink batching behavior.
- commonoptionaluint
max_events
The maximum size of a batch, in events, before it is flushed.
- Default:
20
(events)
- Default:
- commonoptionaluint
timeout_secs
The maximum age of a batch before it is flushed.
- Default:
1
(seconds)
- Default:
- commonoptionalstring
default_namespace
Used as a namespace for metrics that don't have it. A namespace will be prefixed to a metric's name.
- Syntax:
literal
- View examples
- Syntax:
- optionalstring
endpoint
The endpoint to send data to.
- Syntax:
literal
- Only relevant when: region is not set
- View examples
- Syntax:
- commonoptionaltable
healthcheck
Health check options for the sink. See Health checks for more info.
- commonoptionalbool
enabled
Enables/disables the healthcheck upon Vector boot.
- Default:
true
- View examples
- Default:
- optionaltable
request
Configures the sink request behavior.
- optionaltable
adaptive_concurrency
Configure the adaptive concurrency algorithms. These values have been tuned by optimizing simulated results. In general you should not need to adjust these.
- optionalfloat
decrease_ratio
The fraction of the current value to set the new concurrency limit when decreasing the limit. Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases. Note that the new limit is rounded down after applying this ratio.
- Default:
0.9
- Default:
- optionalfloat
ewma_alpha
The adaptive concurrency algorithm uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. This value controls how heavily new measurements are weighted compared to older ones. Valid values are greater than 0 and less than 1. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.
- Default:
0.7
- Default:
- optionalfloat
rtt_threshold_ratio
When comparing the past RTT average to the current measurements, we ignore changes that are less than this ratio higher than the past RTT. Valid values are greater than or equal to 0. Larger values cause the algorithm to ignore larger increases in the RTT.
- Default:
0.05
- Default:
- commonoptionaluint
concurrency
The maximum number of in-flight requests allowed at any given time, or "adaptive" to allow Vector to automatically set the limit based on current network and service conditions.
- Default:
5
(requests)
- Default:
- commonoptionaluint
rate_limit_duration_secs
The time window, in seconds, used for the
rate_limit_num
option.- Default:
1
(seconds)
- Default:
- commonoptionaluint
rate_limit_num
The maximum number of requests allowed within the
rate_limit_duration_secs
time window.- Default:
5
- Default:
- optionaluint
retry_attempts
The maximum number of retries to make for failed requests. The default, for all intents and purposes, represents an infinite number of retries.
- Default:
18446744073709552000
- Default:
- optionaluint
retry_initial_backoff_secs
The amount of time to wait before attempting the first retry for a failed request. Once, the first retry has failed the fibonacci sequence will be used to select future backoffs.
- Default:
1
(seconds)
- Default:
- optionaluint
retry_max_duration_secs
The maximum amount of time, in seconds, to wait between retries.
- Default:
10
(seconds)
- Default:
- commonoptionaluint
timeout_secs
The maximum time a request can take before being aborted. It is highly recommended that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.
- Default:
60
(seconds)
- Default:
Telemetry
This component provides the following metrics that can be retrieved through
the internal_metrics
source. See the
metrics section in the
monitoring page for more info.
- counter
events_in_total
The total number of events accepted by this component. This metric includes the following tags:
component_kind
- The Vector component kind.component_name
- The Vector component ID.component_type
- The Vector component type.instance
- The Vector instance identified by host and port.job
- The name of the job producing Vector metrics.
- counter
events_out_total
The total number of events emitted by this component. This metric includes the following tags:
component_kind
- The Vector component kind.component_name
- The Vector component ID.component_type
- The Vector component type.instance
- The Vector instance identified by host and port.job
- The name of the job producing Vector metrics.
How It Works
Health checks
Health checks ensure that the downstream service is accessible and ready to accept data. This check is performed upon sink initialization. If the health check fails an error will be logged and Vector will proceed to start.
Require health checks
If you'd like to exit immediately upon a health
check failure, you can pass the
--require-healthy
flag:
vector --config /etc/vector/vector.toml --require-healthy
Disable health checks
If you'd like to disable health checks for this
sink you can set the healthcheck
option to
false
.
Partitioning
Vector supports dynamic configuration values through a simple template syntax. If an option supports templating, it will be noted with a badge and you can use event fields to create dynamic values. For example:
[sinks.my-sink]dynamic_option = "application={{ application_id }}"
In the above example, the application_id
for each event will be
used to partition outgoing data.
Rate limits & adapative concurrency
Adaptive Request Concurrency (ARC)
Adaptive Requst Concurrency is a feature of Vector that does away with static rate limits and automatically optimizes HTTP concurrency limits based on downstream service responses. The underlying mechanism is a feedback loop inspired by TCP congestion control algorithms. Checkout the announcement blog post,
We highly recommend enabling this feature as it improves performance and reliability of Vector and the systems it communicates with.
To enable, set the request.concurrency
option to adaptive
:
[sinks.my-sink]request.concurrency = "adaptive"
Static rate limits
If Adaptive Request Concurrency is not for you, you can manually
set static rate limits with the request.rate_limit_duration_secs
,
request.rate_limit_num
, and request.concurrency
options:
[sinks.my-sink]request.rate_limit_duration_secs = 1request.rate_limit_num = 10request.concurrency = 10
Retry policy
Vector will retry failed requests (status == 429, >= 500, and != 501).
Other responses will not be retried. You can control the number of
retry attempts and backoff rate with the request.retry_attempts
and
request.retry_backoff_secs
options.
State
This component is stateless, meaning its behavior is consistent across each input.